var/home/core/zuul-output/0000755000175000017500000000000015110526336014527 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015110535257015475 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004551532715110535252017710 0ustar rootrootNov 23 06:44:57 crc systemd[1]: Starting Kubernetes Kubelet... Nov 23 06:44:57 crc restorecon[4558]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 06:44:57 crc restorecon[4558]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 06:44:57 crc restorecon[4558]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 23 06:44:58 crc kubenswrapper[4559]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 23 06:44:58 crc kubenswrapper[4559]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 23 06:44:58 crc kubenswrapper[4559]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 23 06:44:58 crc kubenswrapper[4559]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 23 06:44:58 crc kubenswrapper[4559]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 23 06:44:58 crc kubenswrapper[4559]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.152619 4559 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155519 4559 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155546 4559 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155558 4559 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155563 4559 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155567 4559 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155571 4559 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155574 4559 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155578 4559 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155582 4559 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155586 4559 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155589 4559 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155592 4559 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155596 4559 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155601 4559 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155605 4559 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155608 4559 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155611 4559 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155614 4559 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155618 4559 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155621 4559 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155624 4559 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155627 4559 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155630 4559 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155634 4559 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155650 4559 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155655 4559 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155659 4559 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155664 4559 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155668 4559 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155673 4559 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155677 4559 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155682 4559 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155686 4559 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155690 4559 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155693 4559 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155697 4559 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155700 4559 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155704 4559 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155708 4559 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155711 4559 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155715 4559 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155718 4559 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155722 4559 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155725 4559 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155728 4559 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155732 4559 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155735 4559 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155738 4559 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155742 4559 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155745 4559 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155748 4559 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155751 4559 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155755 4559 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155758 4559 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155761 4559 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155764 4559 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155768 4559 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155772 4559 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155775 4559 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155778 4559 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155781 4559 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155784 4559 feature_gate.go:330] unrecognized feature gate: Example Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155787 4559 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155792 4559 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155796 4559 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155799 4559 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155802 4559 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155805 4559 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155808 4559 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155811 4559 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.155814 4559 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156330 4559 flags.go:64] FLAG: --address="0.0.0.0" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156340 4559 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156348 4559 flags.go:64] FLAG: --anonymous-auth="true" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156353 4559 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156357 4559 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156361 4559 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156366 4559 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156370 4559 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156374 4559 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156378 4559 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156382 4559 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156386 4559 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156390 4559 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156394 4559 flags.go:64] FLAG: --cgroup-root="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156398 4559 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156402 4559 flags.go:64] FLAG: --client-ca-file="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156406 4559 flags.go:64] FLAG: --cloud-config="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156410 4559 flags.go:64] FLAG: --cloud-provider="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156413 4559 flags.go:64] FLAG: --cluster-dns="[]" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156417 4559 flags.go:64] FLAG: --cluster-domain="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156421 4559 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156424 4559 flags.go:64] FLAG: --config-dir="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156428 4559 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156432 4559 flags.go:64] FLAG: --container-log-max-files="5" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156437 4559 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156441 4559 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156445 4559 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156449 4559 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156452 4559 flags.go:64] FLAG: --contention-profiling="false" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156456 4559 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156459 4559 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156463 4559 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156467 4559 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156472 4559 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156475 4559 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156479 4559 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156482 4559 flags.go:64] FLAG: --enable-load-reader="false" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156486 4559 flags.go:64] FLAG: --enable-server="true" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156490 4559 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156495 4559 flags.go:64] FLAG: --event-burst="100" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156498 4559 flags.go:64] FLAG: --event-qps="50" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156502 4559 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156506 4559 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156509 4559 flags.go:64] FLAG: --eviction-hard="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156514 4559 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156517 4559 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156521 4559 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156525 4559 flags.go:64] FLAG: --eviction-soft="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156539 4559 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156542 4559 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156546 4559 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156550 4559 flags.go:64] FLAG: --experimental-mounter-path="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156554 4559 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156558 4559 flags.go:64] FLAG: --fail-swap-on="true" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156561 4559 flags.go:64] FLAG: --feature-gates="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156565 4559 flags.go:64] FLAG: --file-check-frequency="20s" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156571 4559 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156575 4559 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156579 4559 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156582 4559 flags.go:64] FLAG: --healthz-port="10248" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156586 4559 flags.go:64] FLAG: --help="false" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156590 4559 flags.go:64] FLAG: --hostname-override="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156593 4559 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156597 4559 flags.go:64] FLAG: --http-check-frequency="20s" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156601 4559 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156604 4559 flags.go:64] FLAG: --image-credential-provider-config="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156608 4559 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156612 4559 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156616 4559 flags.go:64] FLAG: --image-service-endpoint="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156620 4559 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156623 4559 flags.go:64] FLAG: --kube-api-burst="100" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156627 4559 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156631 4559 flags.go:64] FLAG: --kube-api-qps="50" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156635 4559 flags.go:64] FLAG: --kube-reserved="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156650 4559 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156655 4559 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156658 4559 flags.go:64] FLAG: --kubelet-cgroups="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156662 4559 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156666 4559 flags.go:64] FLAG: --lock-file="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156669 4559 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156673 4559 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156677 4559 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156682 4559 flags.go:64] FLAG: --log-json-split-stream="false" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156686 4559 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156689 4559 flags.go:64] FLAG: --log-text-split-stream="false" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156693 4559 flags.go:64] FLAG: --logging-format="text" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156697 4559 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156701 4559 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156704 4559 flags.go:64] FLAG: --manifest-url="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156708 4559 flags.go:64] FLAG: --manifest-url-header="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156713 4559 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156716 4559 flags.go:64] FLAG: --max-open-files="1000000" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156721 4559 flags.go:64] FLAG: --max-pods="110" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156724 4559 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156729 4559 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156732 4559 flags.go:64] FLAG: --memory-manager-policy="None" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156736 4559 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156739 4559 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156743 4559 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156746 4559 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156754 4559 flags.go:64] FLAG: --node-status-max-images="50" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156758 4559 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156762 4559 flags.go:64] FLAG: --oom-score-adj="-999" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156765 4559 flags.go:64] FLAG: --pod-cidr="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156769 4559 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156775 4559 flags.go:64] FLAG: --pod-manifest-path="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156778 4559 flags.go:64] FLAG: --pod-max-pids="-1" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156782 4559 flags.go:64] FLAG: --pods-per-core="0" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156786 4559 flags.go:64] FLAG: --port="10250" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156789 4559 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156794 4559 flags.go:64] FLAG: --provider-id="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156798 4559 flags.go:64] FLAG: --qos-reserved="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156802 4559 flags.go:64] FLAG: --read-only-port="10255" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156805 4559 flags.go:64] FLAG: --register-node="true" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156809 4559 flags.go:64] FLAG: --register-schedulable="true" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156813 4559 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156819 4559 flags.go:64] FLAG: --registry-burst="10" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156823 4559 flags.go:64] FLAG: --registry-qps="5" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156827 4559 flags.go:64] FLAG: --reserved-cpus="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156831 4559 flags.go:64] FLAG: --reserved-memory="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156835 4559 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156841 4559 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156845 4559 flags.go:64] FLAG: --rotate-certificates="false" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156849 4559 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156853 4559 flags.go:64] FLAG: --runonce="false" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156856 4559 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156860 4559 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156864 4559 flags.go:64] FLAG: --seccomp-default="false" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156868 4559 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156872 4559 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156876 4559 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156880 4559 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156883 4559 flags.go:64] FLAG: --storage-driver-password="root" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156887 4559 flags.go:64] FLAG: --storage-driver-secure="false" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156891 4559 flags.go:64] FLAG: --storage-driver-table="stats" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156895 4559 flags.go:64] FLAG: --storage-driver-user="root" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156898 4559 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156902 4559 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156906 4559 flags.go:64] FLAG: --system-cgroups="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156909 4559 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156915 4559 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156918 4559 flags.go:64] FLAG: --tls-cert-file="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156922 4559 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156926 4559 flags.go:64] FLAG: --tls-min-version="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156929 4559 flags.go:64] FLAG: --tls-private-key-file="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156933 4559 flags.go:64] FLAG: --topology-manager-policy="none" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156936 4559 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156940 4559 flags.go:64] FLAG: --topology-manager-scope="container" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156944 4559 flags.go:64] FLAG: --v="2" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156948 4559 flags.go:64] FLAG: --version="false" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156953 4559 flags.go:64] FLAG: --vmodule="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156958 4559 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.156961 4559 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157049 4559 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157054 4559 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157057 4559 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157061 4559 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157064 4559 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157068 4559 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157072 4559 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157075 4559 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157079 4559 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157083 4559 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157087 4559 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157090 4559 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157094 4559 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157098 4559 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157102 4559 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157106 4559 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157110 4559 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157114 4559 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157117 4559 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157120 4559 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157123 4559 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157127 4559 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157130 4559 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157133 4559 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157136 4559 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157139 4559 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157143 4559 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157146 4559 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157149 4559 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157152 4559 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157156 4559 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157159 4559 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157165 4559 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157168 4559 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157171 4559 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157174 4559 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157177 4559 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157181 4559 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157184 4559 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157188 4559 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157191 4559 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157195 4559 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157198 4559 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157201 4559 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157204 4559 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157207 4559 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157211 4559 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157214 4559 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157218 4559 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157221 4559 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157224 4559 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157228 4559 feature_gate.go:330] unrecognized feature gate: Example Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157232 4559 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157236 4559 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157240 4559 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157245 4559 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157248 4559 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157252 4559 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157256 4559 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157259 4559 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157263 4559 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157266 4559 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157269 4559 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157273 4559 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157277 4559 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157280 4559 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157286 4559 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157290 4559 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157294 4559 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157297 4559 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.157300 4559 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.157310 4559 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.165157 4559 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.165191 4559 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165257 4559 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165270 4559 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165275 4559 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165280 4559 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165285 4559 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165289 4559 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165293 4559 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165297 4559 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165302 4559 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165307 4559 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165311 4559 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165314 4559 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165318 4559 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165321 4559 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165326 4559 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165330 4559 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165335 4559 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165339 4559 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165344 4559 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165348 4559 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165353 4559 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165357 4559 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165360 4559 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165364 4559 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165367 4559 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165371 4559 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165374 4559 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165396 4559 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165400 4559 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165403 4559 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165407 4559 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165410 4559 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165414 4559 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165418 4559 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165421 4559 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165425 4559 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165428 4559 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165432 4559 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165436 4559 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165439 4559 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165443 4559 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165446 4559 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165450 4559 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165453 4559 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165458 4559 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165464 4559 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165468 4559 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165472 4559 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165475 4559 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165479 4559 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165483 4559 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165487 4559 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165491 4559 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165495 4559 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165498 4559 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165501 4559 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165505 4559 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165509 4559 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165512 4559 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165516 4559 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165520 4559 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165524 4559 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165536 4559 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165539 4559 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165542 4559 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165546 4559 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165551 4559 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165555 4559 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165559 4559 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165563 4559 feature_gate.go:330] unrecognized feature gate: Example Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165566 4559 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.165573 4559 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165693 4559 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165701 4559 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165705 4559 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165725 4559 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165730 4559 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165734 4559 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165738 4559 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165742 4559 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165747 4559 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165751 4559 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165754 4559 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165758 4559 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165761 4559 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165765 4559 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165768 4559 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165772 4559 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165776 4559 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165780 4559 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165784 4559 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165787 4559 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165791 4559 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165794 4559 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165799 4559 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165802 4559 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165806 4559 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165810 4559 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165813 4559 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165816 4559 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165820 4559 feature_gate.go:330] unrecognized feature gate: Example Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165823 4559 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165826 4559 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165829 4559 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165832 4559 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165835 4559 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165838 4559 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165842 4559 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165845 4559 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165849 4559 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165853 4559 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165856 4559 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165859 4559 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165862 4559 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165866 4559 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165870 4559 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165874 4559 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165878 4559 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165881 4559 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165885 4559 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165888 4559 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165891 4559 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165894 4559 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165898 4559 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165901 4559 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165904 4559 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165907 4559 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165910 4559 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165914 4559 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165917 4559 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165920 4559 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165923 4559 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165927 4559 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165931 4559 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165935 4559 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165939 4559 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165942 4559 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165945 4559 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165949 4559 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165952 4559 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165955 4559 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165959 4559 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.165962 4559 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.165967 4559 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.166422 4559 server.go:940] "Client rotation is on, will bootstrap in background" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.169036 4559 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.169113 4559 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.170103 4559 server.go:997] "Starting client certificate rotation" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.170131 4559 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.170549 4559 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-17 23:16:36.006415313 +0000 UTC Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.170604 4559 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 592h31m37.835812816s for next certificate rotation Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.182001 4559 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.183259 4559 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.192729 4559 log.go:25] "Validated CRI v1 runtime API" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.209475 4559 log.go:25] "Validated CRI v1 image API" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.210575 4559 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.213847 4559 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-23-06-41-49-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.213874 4559 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:49 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/containers/storage/overlay-containers/75d81934760b26101869fbd8e4b5954c62b019c1cc3e5a0c9f82ed8de46b3b22/userdata/shm:{mountpoint:/var/lib/containers/storage/overlay-containers/75d81934760b26101869fbd8e4b5954c62b019c1cc3e5a0c9f82ed8de46b3b22/userdata/shm major:0 minor:42 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:50 fsType:tmpfs blockSize:0} overlay_0-43:{mountpoint:/var/lib/containers/storage/overlay/94b752e0a51c0134b00ddef6dc7a933a9d7c1d9bdc88a18dae4192a0d557d623/merged major:0 minor:43 fsType:overlay blockSize:0}] Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.227405 4559 manager.go:217] Machine: {Timestamp:2025-11-23 06:44:58.22580238 +0000 UTC m=+0.247787994 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2445406 MemoryCapacity:33654120448 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:8b1b3dc3-1aa7-4149-9ed7-13994feca6e6 BootID:d5986746-4ff2-4418-9747-f60dccf1cb6d Filesystems:[{Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true} {Device:overlay_0-43 DeviceMajor:0 DeviceMinor:43 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:49 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:50 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/var/lib/containers/storage/overlay-containers/75d81934760b26101869fbd8e4b5954c62b019c1cc3e5a0c9f82ed8de46b3b22/userdata/shm DeviceMajor:0 DeviceMinor:42 Capacity:65536000 Type:vfs Inodes:4108169 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:07:2f:89 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:enp3s0 MacAddress:fa:16:3e:07:2f:89 Speed:-1 Mtu:1500} {Name:enp7s0 MacAddress:fa:16:3e:75:7f:22 Speed:-1 Mtu:1440} {Name:enp7s0.20 MacAddress:52:54:00:4f:84:a8 Speed:-1 Mtu:1436} {Name:enp7s0.21 MacAddress:52:54:00:e4:d9:64 Speed:-1 Mtu:1436} {Name:enp7s0.22 MacAddress:52:54:00:63:9b:3c Speed:-1 Mtu:1436} {Name:eth10 MacAddress:22:14:16:0e:38:0e Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:62:34:21:e2:f5:8f Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654120448 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:65536 Type:Data Level:1} {Id:0 Size:65536 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:65536 Type:Data Level:1} {Id:1 Size:65536 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:65536 Type:Data Level:1} {Id:10 Size:65536 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:65536 Type:Data Level:1} {Id:11 Size:65536 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:65536 Type:Data Level:1} {Id:2 Size:65536 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:65536 Type:Data Level:1} {Id:3 Size:65536 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:65536 Type:Data Level:1} {Id:4 Size:65536 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:65536 Type:Data Level:1} {Id:5 Size:65536 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:65536 Type:Data Level:1} {Id:6 Size:65536 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:65536 Type:Data Level:1} {Id:7 Size:65536 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:65536 Type:Data Level:1} {Id:8 Size:65536 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:65536 Type:Data Level:1} {Id:9 Size:65536 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.227567 4559 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.227702 4559 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.228232 4559 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.228385 4559 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.228402 4559 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.228548 4559 topology_manager.go:138] "Creating topology manager with none policy" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.228556 4559 container_manager_linux.go:303] "Creating device plugin manager" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.228852 4559 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.228882 4559 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.229368 4559 state_mem.go:36] "Initialized new in-memory state store" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.229433 4559 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.231049 4559 kubelet.go:418] "Attempting to sync node with API server" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.231065 4559 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.231085 4559 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.231094 4559 kubelet.go:324] "Adding apiserver pod source" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.231106 4559 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.233366 4559 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.233891 4559 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.235009 4559 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 192.168.25.152:6443: connect: connection refused Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.235010 4559 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 192.168.25.152:6443: connect: connection refused Nov 23 06:44:58 crc kubenswrapper[4559]: E1123 06:44:58.235113 4559 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 192.168.25.152:6443: connect: connection refused" logger="UnhandledError" Nov 23 06:44:58 crc kubenswrapper[4559]: E1123 06:44:58.235130 4559 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 192.168.25.152:6443: connect: connection refused" logger="UnhandledError" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.235350 4559 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.236187 4559 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.236210 4559 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.236218 4559 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.236225 4559 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.236236 4559 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.236243 4559 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.236249 4559 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.236259 4559 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.236267 4559 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.236275 4559 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.236285 4559 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.236293 4559 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.236865 4559 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.237136 4559 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 192.168.25.152:6443: connect: connection refused Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.237205 4559 server.go:1280] "Started kubelet" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.237593 4559 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.237601 4559 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 23 06:44:58 crc systemd[1]: Started Kubernetes Kubelet. Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.238519 4559 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.239191 4559 server.go:460] "Adding debug handlers to kubelet server" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.239238 4559 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.239262 4559 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.239523 4559 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 06:01:48.117048896 +0000 UTC Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.243635 4559 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.244240 4559 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 23 06:44:58 crc kubenswrapper[4559]: E1123 06:44:58.244212 4559 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.244276 4559 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.244850 4559 factory.go:55] Registering systemd factory Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.244864 4559 factory.go:221] Registration of the systemd container factory successfully Nov 23 06:44:58 crc kubenswrapper[4559]: E1123 06:44:58.244887 4559 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.25.152:6443: connect: connection refused" interval="200ms" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.245151 4559 factory.go:153] Registering CRI-O factory Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.245170 4559 factory.go:221] Registration of the crio container factory successfully Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.245179 4559 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 192.168.25.152:6443: connect: connection refused Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.245224 4559 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 23 06:44:58 crc kubenswrapper[4559]: E1123 06:44:58.245234 4559 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 192.168.25.152:6443: connect: connection refused" logger="UnhandledError" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.245244 4559 factory.go:103] Registering Raw factory Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.245259 4559 manager.go:1196] Started watching for new ooms in manager Nov 23 06:44:58 crc kubenswrapper[4559]: E1123 06:44:58.244759 4559 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 192.168.25.152:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187a8fc92abcbfc8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-23 06:44:58.237181896 +0000 UTC m=+0.259167510,LastTimestamp:2025-11-23 06:44:58.237181896 +0000 UTC m=+0.259167510,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.245800 4559 manager.go:319] Starting recovery of all containers Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251584 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251615 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251628 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251637 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251662 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251672 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251681 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251689 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251701 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251709 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251718 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251726 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251736 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251748 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251756 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251763 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251773 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251782 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251789 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251798 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251805 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251813 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251821 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251830 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251837 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251848 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251858 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251884 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251894 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251902 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251909 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251917 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251929 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251938 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251946 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251955 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251963 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251972 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.251981 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.252002 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.252010 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.252018 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.252027 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.252055 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.252065 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.252074 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.252082 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.252090 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.252099 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.252108 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.252117 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.252126 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.252138 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.252147 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.252157 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.252166 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.252175 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.252183 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.252191 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253244 4559 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253270 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253282 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253291 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253298 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253307 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253315 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253323 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253332 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253340 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253349 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253356 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253364 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253371 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253394 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253402 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253410 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253418 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253425 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253433 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253442 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253450 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253458 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253467 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253475 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253483 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253493 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253743 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253754 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253762 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253771 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253779 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253787 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253796 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253804 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253812 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253820 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253829 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253836 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253882 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253895 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253903 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253911 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253919 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253926 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253935 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253948 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253958 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253967 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.253976 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.254008 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.254018 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.254027 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.254037 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.254045 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.254220 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.254414 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.254430 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.254439 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.254449 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.254458 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.254466 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.254473 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.254482 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.254489 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.254497 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.254506 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.254513 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.254952 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.256631 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.256686 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.256698 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.256717 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.256738 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.256770 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.256780 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.256788 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257356 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257433 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257449 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257466 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257477 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257487 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257497 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257521 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257546 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257555 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257563 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257573 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257583 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257591 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257599 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257607 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257615 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257624 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257633 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257658 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257667 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257675 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257683 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257691 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257699 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257709 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257716 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257726 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257733 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257742 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257750 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257758 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257766 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257774 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257781 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257789 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257797 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257806 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257813 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257821 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257830 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257837 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257846 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257855 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257862 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257871 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257879 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257888 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257897 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257906 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257915 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257924 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257932 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257942 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257951 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257958 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257967 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.257974 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.258012 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.258022 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.258029 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.258037 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.258047 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.258055 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.258063 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.258070 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.258077 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.258085 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.258094 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.258103 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.258111 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.258119 4559 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.258129 4559 reconstruct.go:97] "Volume reconstruction finished" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.258135 4559 reconciler.go:26] "Reconciler: start to sync state" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.259172 4559 manager.go:324] Recovery completed Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.266007 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.267156 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.267190 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.267200 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.267827 4559 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.267837 4559 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.267856 4559 state_mem.go:36] "Initialized new in-memory state store" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.271069 4559 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.272336 4559 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.272365 4559 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.272383 4559 kubelet.go:2335] "Starting kubelet main sync loop" Nov 23 06:44:58 crc kubenswrapper[4559]: E1123 06:44:58.272418 4559 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.272628 4559 policy_none.go:49] "None policy: Start" Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.272814 4559 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 192.168.25.152:6443: connect: connection refused Nov 23 06:44:58 crc kubenswrapper[4559]: E1123 06:44:58.272866 4559 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 192.168.25.152:6443: connect: connection refused" logger="UnhandledError" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.275052 4559 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.275080 4559 state_mem.go:35] "Initializing new in-memory state store" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.316359 4559 manager.go:334] "Starting Device Plugin manager" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.316390 4559 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.316401 4559 server.go:79] "Starting device plugin registration server" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.316683 4559 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.316698 4559 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.316812 4559 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.316878 4559 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.316889 4559 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 23 06:44:58 crc kubenswrapper[4559]: E1123 06:44:58.321558 4559 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.373073 4559 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.373133 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.373802 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.373829 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.373838 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.373921 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.374068 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.374105 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.374439 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.374470 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.374481 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.374601 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.374728 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.374757 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.374757 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.374798 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.374768 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.375263 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.375303 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.375312 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.375404 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.375819 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.375847 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.376096 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.376140 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.376150 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.376244 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.376333 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.376349 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.376366 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.376371 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.376378 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.376767 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.376790 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.376800 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.376917 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.376939 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.376948 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.377387 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.377433 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.377443 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.377596 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.377621 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.378382 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.378407 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.378418 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.416946 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.417504 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.417539 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.417549 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.417563 4559 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 23 06:44:58 crc kubenswrapper[4559]: E1123 06:44:58.417875 4559 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 192.168.25.152:6443: connect: connection refused" node="crc" Nov 23 06:44:58 crc kubenswrapper[4559]: E1123 06:44:58.445450 4559 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.25.152:6443: connect: connection refused" interval="400ms" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.459623 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.459660 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.459678 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.459705 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.459735 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.459757 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.459774 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.459801 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.459816 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.459839 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.459857 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.459882 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.459934 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.459965 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.459989 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.560741 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.560781 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.560797 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.560812 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.560826 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.560859 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.560871 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.560887 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.560902 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.560903 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.560919 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.560921 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.560936 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.560946 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.560953 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.560967 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.560969 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.560979 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.560871 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.560890 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.561031 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.561039 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.560996 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.561088 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.560978 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.561111 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.560929 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.561124 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.561132 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.561162 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.618844 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.619922 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.619951 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.619982 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.620214 4559 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 23 06:44:58 crc kubenswrapper[4559]: E1123 06:44:58.620503 4559 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 192.168.25.152:6443: connect: connection refused" node="crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.705070 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.722951 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.727156 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-d42922348c8abaaf10ac86412b84900984ff7433be811ad9a4515918b7ed8a27 WatchSource:0}: Error finding container d42922348c8abaaf10ac86412b84900984ff7433be811ad9a4515918b7ed8a27: Status 404 returned error can't find the container with id d42922348c8abaaf10ac86412b84900984ff7433be811ad9a4515918b7ed8a27 Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.735112 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-87809ecc9a6655e6761ba3f76559ce7be7c6f6fa25cfc2879b0a45ed6485d66c WatchSource:0}: Error finding container 87809ecc9a6655e6761ba3f76559ce7be7c6f6fa25cfc2879b0a45ed6485d66c: Status 404 returned error can't find the container with id 87809ecc9a6655e6761ba3f76559ce7be7c6f6fa25cfc2879b0a45ed6485d66c Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.735295 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.741426 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: I1123 06:44:58.745219 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.748230 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-c0a697859959ce030c0139c770a643814b0144e97286013d52fa64b2d02003af WatchSource:0}: Error finding container c0a697859959ce030c0139c770a643814b0144e97286013d52fa64b2d02003af: Status 404 returned error can't find the container with id c0a697859959ce030c0139c770a643814b0144e97286013d52fa64b2d02003af Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.755584 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-8a5dbaa01d60c7266f4feb35583cc15ceb6b03efeac1c9b9920e47a119e26f96 WatchSource:0}: Error finding container 8a5dbaa01d60c7266f4feb35583cc15ceb6b03efeac1c9b9920e47a119e26f96: Status 404 returned error can't find the container with id 8a5dbaa01d60c7266f4feb35583cc15ceb6b03efeac1c9b9920e47a119e26f96 Nov 23 06:44:58 crc kubenswrapper[4559]: W1123 06:44:58.758509 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-6bbc6e0282b08f25997f7bf23906074efb99e4e207ac4b70474e74d32a0d5ffe WatchSource:0}: Error finding container 6bbc6e0282b08f25997f7bf23906074efb99e4e207ac4b70474e74d32a0d5ffe: Status 404 returned error can't find the container with id 6bbc6e0282b08f25997f7bf23906074efb99e4e207ac4b70474e74d32a0d5ffe Nov 23 06:44:58 crc kubenswrapper[4559]: E1123 06:44:58.846307 4559 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.25.152:6443: connect: connection refused" interval="800ms" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.021477 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.022680 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.022711 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.022722 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.022743 4559 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 23 06:44:59 crc kubenswrapper[4559]: E1123 06:44:59.023124 4559 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 192.168.25.152:6443: connect: connection refused" node="crc" Nov 23 06:44:59 crc kubenswrapper[4559]: W1123 06:44:59.085869 4559 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 192.168.25.152:6443: connect: connection refused Nov 23 06:44:59 crc kubenswrapper[4559]: E1123 06:44:59.085931 4559 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 192.168.25.152:6443: connect: connection refused" logger="UnhandledError" Nov 23 06:44:59 crc kubenswrapper[4559]: W1123 06:44:59.166620 4559 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 192.168.25.152:6443: connect: connection refused Nov 23 06:44:59 crc kubenswrapper[4559]: E1123 06:44:59.166700 4559 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 192.168.25.152:6443: connect: connection refused" logger="UnhandledError" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.238225 4559 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 192.168.25.152:6443: connect: connection refused Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.240283 4559 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 14:02:20.797689729 +0000 UTC Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.240330 4559 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1207h17m21.557361769s for next certificate rotation Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.277533 4559 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="a415a92a899290d1b121b3ba83e3d3ca0a84a001256e242cac011a9933fc1cd9" exitCode=0 Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.277608 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"a415a92a899290d1b121b3ba83e3d3ca0a84a001256e242cac011a9933fc1cd9"} Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.277697 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6bbc6e0282b08f25997f7bf23906074efb99e4e207ac4b70474e74d32a0d5ffe"} Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.277766 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.278382 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.278417 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.278427 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.278931 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c"} Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.278976 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8a5dbaa01d60c7266f4feb35583cc15ceb6b03efeac1c9b9920e47a119e26f96"} Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.280171 4559 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6" exitCode=0 Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.280249 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6"} Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.280283 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c0a697859959ce030c0139c770a643814b0144e97286013d52fa64b2d02003af"} Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.280359 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.281155 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.281183 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.281192 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.281694 4559 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="4119aa9bf32e07b601b8427be089972a6231e9441d7ba5974f3f29e3fb434345" exitCode=0 Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.281745 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"4119aa9bf32e07b601b8427be089972a6231e9441d7ba5974f3f29e3fb434345"} Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.281762 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"87809ecc9a6655e6761ba3f76559ce7be7c6f6fa25cfc2879b0a45ed6485d66c"} Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.281822 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.282319 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.282506 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.282526 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.282534 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.282780 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.282809 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.282817 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.283566 4559 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="593cf113905c89a79b8db618d0011376eb1db9755e6580e56c16b3af3c358675" exitCode=0 Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.283595 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"593cf113905c89a79b8db618d0011376eb1db9755e6580e56c16b3af3c358675"} Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.283612 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"d42922348c8abaaf10ac86412b84900984ff7433be811ad9a4515918b7ed8a27"} Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.283805 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.286279 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.286310 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.286320 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:44:59 crc kubenswrapper[4559]: W1123 06:44:59.602549 4559 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 192.168.25.152:6443: connect: connection refused Nov 23 06:44:59 crc kubenswrapper[4559]: E1123 06:44:59.602615 4559 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 192.168.25.152:6443: connect: connection refused" logger="UnhandledError" Nov 23 06:44:59 crc kubenswrapper[4559]: E1123 06:44:59.647906 4559 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.25.152:6443: connect: connection refused" interval="1.6s" Nov 23 06:44:59 crc kubenswrapper[4559]: W1123 06:44:59.683920 4559 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 192.168.25.152:6443: connect: connection refused Nov 23 06:44:59 crc kubenswrapper[4559]: E1123 06:44:59.683984 4559 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 192.168.25.152:6443: connect: connection refused" logger="UnhandledError" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.824005 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.824987 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.825014 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.825024 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:44:59 crc kubenswrapper[4559]: I1123 06:44:59.825043 4559 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 23 06:44:59 crc kubenswrapper[4559]: E1123 06:44:59.825424 4559 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 192.168.25.152:6443: connect: connection refused" node="crc" Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.287315 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"411e9c23598439e930509a324a28db67bc31bad2613822349d7a2fed97937e96"} Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.287568 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"51539e1fed5ea7e5899c119f4a93fd819aaba7ce335225a20b766a4dcbcfcad4"} Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.287580 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a7cceb63202f93e3545d380835e59368e9c8dfb8f818cf6753fd2790574885e9"} Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.287676 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.288250 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.288283 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.288293 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.289844 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7"} Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.289883 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d"} Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.289895 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5"} Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.289859 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.290467 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.290488 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.290496 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.292093 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00"} Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.292115 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c"} Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.292125 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea"} Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.292136 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59"} Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.292145 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b"} Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.292206 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.292726 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.292753 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.292761 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.293410 4559 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="69200ad696a1b3bf10a8fda564ab6d301191dd0ad30782d5bf2ba01da2a468bf" exitCode=0 Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.293439 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"69200ad696a1b3bf10a8fda564ab6d301191dd0ad30782d5bf2ba01da2a468bf"} Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.293536 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.294052 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.294073 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.294081 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.294590 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"911f55a9468868f66060d4080ebf586aa360927045d4917667d2cb186aac82fa"} Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.294656 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.295198 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.295234 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.295245 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:00 crc kubenswrapper[4559]: I1123 06:45:00.533606 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:45:01 crc kubenswrapper[4559]: I1123 06:45:01.298240 4559 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e3fa490a43511f6348a17eb174d75ab2c27014a88ed743bed4bf9f1f8e23c283" exitCode=0 Nov 23 06:45:01 crc kubenswrapper[4559]: I1123 06:45:01.298319 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:45:01 crc kubenswrapper[4559]: I1123 06:45:01.298326 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e3fa490a43511f6348a17eb174d75ab2c27014a88ed743bed4bf9f1f8e23c283"} Nov 23 06:45:01 crc kubenswrapper[4559]: I1123 06:45:01.298406 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:45:01 crc kubenswrapper[4559]: I1123 06:45:01.298428 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:45:01 crc kubenswrapper[4559]: I1123 06:45:01.298909 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:01 crc kubenswrapper[4559]: I1123 06:45:01.298932 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:01 crc kubenswrapper[4559]: I1123 06:45:01.298941 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:01 crc kubenswrapper[4559]: I1123 06:45:01.299143 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:01 crc kubenswrapper[4559]: I1123 06:45:01.299180 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:01 crc kubenswrapper[4559]: I1123 06:45:01.299190 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:01 crc kubenswrapper[4559]: I1123 06:45:01.299204 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:01 crc kubenswrapper[4559]: I1123 06:45:01.299219 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:01 crc kubenswrapper[4559]: I1123 06:45:01.299227 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:01 crc kubenswrapper[4559]: I1123 06:45:01.426462 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:45:01 crc kubenswrapper[4559]: I1123 06:45:01.427206 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:01 crc kubenswrapper[4559]: I1123 06:45:01.427235 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:01 crc kubenswrapper[4559]: I1123 06:45:01.427245 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:01 crc kubenswrapper[4559]: I1123 06:45:01.427282 4559 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 23 06:45:02 crc kubenswrapper[4559]: I1123 06:45:02.302311 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c22e5a020b14f0efb3f405f8213d93fc3f54fc72f9f6c9c759488ec989a404de"} Nov 23 06:45:02 crc kubenswrapper[4559]: I1123 06:45:02.302348 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"19154f6b2e570f1c6dc9a60075e82b244c6f7c33cc282006417d32614879bf4c"} Nov 23 06:45:02 crc kubenswrapper[4559]: I1123 06:45:02.302359 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f890d46893102d69aa5c9dd979d26d62285e316449eb8fec8eb7e127c822c8a7"} Nov 23 06:45:02 crc kubenswrapper[4559]: I1123 06:45:02.302368 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b99b435654bd4c91c8b0b834591d8731f35b4458f6f325f9f111aff744a857ee"} Nov 23 06:45:02 crc kubenswrapper[4559]: I1123 06:45:02.302368 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:45:02 crc kubenswrapper[4559]: I1123 06:45:02.302415 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:45:02 crc kubenswrapper[4559]: I1123 06:45:02.302775 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"368d28b63dc438ac6b825a671312deba33a0edff83fb939945d1d4ac9c1fc1b3"} Nov 23 06:45:02 crc kubenswrapper[4559]: I1123 06:45:02.303003 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:02 crc kubenswrapper[4559]: I1123 06:45:02.303016 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:02 crc kubenswrapper[4559]: I1123 06:45:02.303023 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:02 crc kubenswrapper[4559]: I1123 06:45:02.303030 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:02 crc kubenswrapper[4559]: I1123 06:45:02.303038 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:02 crc kubenswrapper[4559]: I1123 06:45:02.303031 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:03 crc kubenswrapper[4559]: I1123 06:45:03.304786 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:45:03 crc kubenswrapper[4559]: I1123 06:45:03.305354 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:03 crc kubenswrapper[4559]: I1123 06:45:03.305386 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:03 crc kubenswrapper[4559]: I1123 06:45:03.305396 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:03 crc kubenswrapper[4559]: I1123 06:45:03.311097 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 06:45:03 crc kubenswrapper[4559]: I1123 06:45:03.311171 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:45:03 crc kubenswrapper[4559]: I1123 06:45:03.311682 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:03 crc kubenswrapper[4559]: I1123 06:45:03.311706 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:03 crc kubenswrapper[4559]: I1123 06:45:03.311715 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:03 crc kubenswrapper[4559]: I1123 06:45:03.443360 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 23 06:45:03 crc kubenswrapper[4559]: I1123 06:45:03.519784 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:45:03 crc kubenswrapper[4559]: I1123 06:45:03.519906 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:45:03 crc kubenswrapper[4559]: I1123 06:45:03.520008 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:45:03 crc kubenswrapper[4559]: I1123 06:45:03.520728 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:03 crc kubenswrapper[4559]: I1123 06:45:03.520761 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:03 crc kubenswrapper[4559]: I1123 06:45:03.520771 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:03 crc kubenswrapper[4559]: I1123 06:45:03.534132 4559 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 23 06:45:03 crc kubenswrapper[4559]: I1123 06:45:03.534242 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 23 06:45:04 crc kubenswrapper[4559]: I1123 06:45:04.079688 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:45:04 crc kubenswrapper[4559]: I1123 06:45:04.306161 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:45:04 crc kubenswrapper[4559]: I1123 06:45:04.306188 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:45:04 crc kubenswrapper[4559]: I1123 06:45:04.306780 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:04 crc kubenswrapper[4559]: I1123 06:45:04.306810 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:04 crc kubenswrapper[4559]: I1123 06:45:04.306818 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:04 crc kubenswrapper[4559]: I1123 06:45:04.307283 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:04 crc kubenswrapper[4559]: I1123 06:45:04.307300 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:04 crc kubenswrapper[4559]: I1123 06:45:04.307307 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:05 crc kubenswrapper[4559]: I1123 06:45:05.092686 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:45:05 crc kubenswrapper[4559]: I1123 06:45:05.092801 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:45:05 crc kubenswrapper[4559]: I1123 06:45:05.093592 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:05 crc kubenswrapper[4559]: I1123 06:45:05.093618 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:05 crc kubenswrapper[4559]: I1123 06:45:05.093626 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:05 crc kubenswrapper[4559]: I1123 06:45:05.307449 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:45:05 crc kubenswrapper[4559]: I1123 06:45:05.308056 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:05 crc kubenswrapper[4559]: I1123 06:45:05.308082 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:05 crc kubenswrapper[4559]: I1123 06:45:05.308089 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:05 crc kubenswrapper[4559]: I1123 06:45:05.888632 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 23 06:45:05 crc kubenswrapper[4559]: I1123 06:45:05.888883 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:45:05 crc kubenswrapper[4559]: I1123 06:45:05.889946 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:05 crc kubenswrapper[4559]: I1123 06:45:05.889991 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:05 crc kubenswrapper[4559]: I1123 06:45:05.890001 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:08 crc kubenswrapper[4559]: E1123 06:45:08.321610 4559 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 23 06:45:08 crc kubenswrapper[4559]: I1123 06:45:08.626821 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:45:08 crc kubenswrapper[4559]: I1123 06:45:08.627849 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:45:08 crc kubenswrapper[4559]: I1123 06:45:08.629539 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:08 crc kubenswrapper[4559]: I1123 06:45:08.629566 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:08 crc kubenswrapper[4559]: I1123 06:45:08.629579 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:08 crc kubenswrapper[4559]: I1123 06:45:08.631089 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:45:09 crc kubenswrapper[4559]: I1123 06:45:09.124057 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:45:09 crc kubenswrapper[4559]: I1123 06:45:09.127345 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:45:09 crc kubenswrapper[4559]: I1123 06:45:09.315339 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:45:09 crc kubenswrapper[4559]: I1123 06:45:09.316104 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:09 crc kubenswrapper[4559]: I1123 06:45:09.316141 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:09 crc kubenswrapper[4559]: I1123 06:45:09.316150 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:09 crc kubenswrapper[4559]: I1123 06:45:09.742268 4559 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 23 06:45:09 crc kubenswrapper[4559]: I1123 06:45:09.742320 4559 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 23 06:45:10 crc kubenswrapper[4559]: I1123 06:45:10.238786 4559 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 23 06:45:10 crc kubenswrapper[4559]: I1123 06:45:10.259729 4559 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 23 06:45:10 crc kubenswrapper[4559]: I1123 06:45:10.259780 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 23 06:45:10 crc kubenswrapper[4559]: I1123 06:45:10.267337 4559 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 23 06:45:10 crc kubenswrapper[4559]: I1123 06:45:10.267383 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 23 06:45:10 crc kubenswrapper[4559]: I1123 06:45:10.317413 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:45:10 crc kubenswrapper[4559]: I1123 06:45:10.318231 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:10 crc kubenswrapper[4559]: I1123 06:45:10.318265 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:10 crc kubenswrapper[4559]: I1123 06:45:10.318276 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:13 crc kubenswrapper[4559]: I1123 06:45:13.460139 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 23 06:45:13 crc kubenswrapper[4559]: I1123 06:45:13.460238 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:45:13 crc kubenswrapper[4559]: I1123 06:45:13.460959 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:13 crc kubenswrapper[4559]: I1123 06:45:13.460993 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:13 crc kubenswrapper[4559]: I1123 06:45:13.461004 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:13 crc kubenswrapper[4559]: I1123 06:45:13.468050 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 23 06:45:13 crc kubenswrapper[4559]: I1123 06:45:13.523731 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:45:13 crc kubenswrapper[4559]: I1123 06:45:13.523838 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:45:13 crc kubenswrapper[4559]: I1123 06:45:13.524066 4559 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 23 06:45:13 crc kubenswrapper[4559]: I1123 06:45:13.524106 4559 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 23 06:45:13 crc kubenswrapper[4559]: I1123 06:45:13.524463 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:13 crc kubenswrapper[4559]: I1123 06:45:13.524492 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:13 crc kubenswrapper[4559]: I1123 06:45:13.524502 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:13 crc kubenswrapper[4559]: I1123 06:45:13.526941 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:45:13 crc kubenswrapper[4559]: I1123 06:45:13.534926 4559 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 23 06:45:13 crc kubenswrapper[4559]: I1123 06:45:13.534951 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 23 06:45:14 crc kubenswrapper[4559]: I1123 06:45:14.324716 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:45:14 crc kubenswrapper[4559]: I1123 06:45:14.325179 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:45:14 crc kubenswrapper[4559]: I1123 06:45:14.325475 4559 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 23 06:45:14 crc kubenswrapper[4559]: I1123 06:45:14.325515 4559 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 23 06:45:14 crc kubenswrapper[4559]: I1123 06:45:14.326172 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:14 crc kubenswrapper[4559]: I1123 06:45:14.326261 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:14 crc kubenswrapper[4559]: I1123 06:45:14.326330 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:14 crc kubenswrapper[4559]: I1123 06:45:14.326179 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:14 crc kubenswrapper[4559]: I1123 06:45:14.326448 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:14 crc kubenswrapper[4559]: I1123 06:45:14.326470 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:15 crc kubenswrapper[4559]: I1123 06:45:15.262575 4559 trace.go:236] Trace[804551047]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (23-Nov-2025 06:45:01.441) (total time: 13820ms): Nov 23 06:45:15 crc kubenswrapper[4559]: Trace[804551047]: ---"Objects listed" error: 13820ms (06:45:15.262) Nov 23 06:45:15 crc kubenswrapper[4559]: Trace[804551047]: [13.820784318s] [13.820784318s] END Nov 23 06:45:15 crc kubenswrapper[4559]: I1123 06:45:15.262613 4559 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 23 06:45:15 crc kubenswrapper[4559]: I1123 06:45:15.264451 4559 trace.go:236] Trace[158874325]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (23-Nov-2025 06:45:01.753) (total time: 13511ms): Nov 23 06:45:15 crc kubenswrapper[4559]: Trace[158874325]: ---"Objects listed" error: 13511ms (06:45:15.264) Nov 23 06:45:15 crc kubenswrapper[4559]: Trace[158874325]: [13.511142515s] [13.511142515s] END Nov 23 06:45:15 crc kubenswrapper[4559]: I1123 06:45:15.264473 4559 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 23 06:45:15 crc kubenswrapper[4559]: I1123 06:45:15.264503 4559 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 23 06:45:15 crc kubenswrapper[4559]: E1123 06:45:15.264951 4559 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 23 06:45:15 crc kubenswrapper[4559]: I1123 06:45:15.265262 4559 trace.go:236] Trace[303309658]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (23-Nov-2025 06:45:01.612) (total time: 13653ms): Nov 23 06:45:15 crc kubenswrapper[4559]: Trace[303309658]: ---"Objects listed" error: 13652ms (06:45:15.265) Nov 23 06:45:15 crc kubenswrapper[4559]: Trace[303309658]: [13.653063657s] [13.653063657s] END Nov 23 06:45:15 crc kubenswrapper[4559]: I1123 06:45:15.265295 4559 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 23 06:45:15 crc kubenswrapper[4559]: I1123 06:45:15.267556 4559 trace.go:236] Trace[95609059]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (23-Nov-2025 06:45:01.909) (total time: 13357ms): Nov 23 06:45:15 crc kubenswrapper[4559]: Trace[95609059]: ---"Objects listed" error: 13357ms (06:45:15.267) Nov 23 06:45:15 crc kubenswrapper[4559]: Trace[95609059]: [13.357715595s] [13.357715595s] END Nov 23 06:45:15 crc kubenswrapper[4559]: I1123 06:45:15.267572 4559 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 23 06:45:15 crc kubenswrapper[4559]: E1123 06:45:15.269348 4559 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Nov 23 06:45:15 crc kubenswrapper[4559]: I1123 06:45:15.438082 4559 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:33264->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 23 06:45:15 crc kubenswrapper[4559]: I1123 06:45:15.438122 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:33264->192.168.126.11:17697: read: connection reset by peer" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.240328 4559 apiserver.go:52] "Watching apiserver" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.243070 4559 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.243741 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.244047 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.244108 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.244154 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.244242 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.244319 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.244727 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.245621 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.245680 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.245716 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.246990 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.247100 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.247602 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.247669 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.247744 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.247812 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.247860 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.247976 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.249183 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.263000 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.270108 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.270156 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.270176 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.270191 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.270226 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.270242 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.270258 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.270291 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.270307 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.270603 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.270702 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.271131 4559 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.271185 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.275460 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.275632 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.279045 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.280055 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.280078 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.280088 4559 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.280127 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-23 06:45:16.780114993 +0000 UTC m=+18.802100608 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.281920 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.281952 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.281964 4559 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.282014 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-23 06:45:16.781999974 +0000 UTC m=+18.803985589 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.282565 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.283222 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.285472 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.291526 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.298243 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.304553 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.310330 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.331159 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.332856 4559 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00" exitCode=255 Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.332893 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00"} Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.339262 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.339813 4559 scope.go:117] "RemoveContainer" containerID="e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.340346 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.345223 4559 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.346543 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.352663 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.359429 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.368377 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.370603 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.370629 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.370659 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.370676 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.370693 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.370707 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.370722 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.370739 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.370787 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.370806 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.370825 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.370847 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.370866 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.370883 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.370899 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.370898 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.370919 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.370936 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.370953 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.370971 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.370987 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371004 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371022 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371039 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371056 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371073 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371091 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371108 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371124 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371139 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371155 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371171 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371188 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371203 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371221 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371239 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371255 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371271 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371286 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371302 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371318 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371342 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371360 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371378 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371396 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371412 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371428 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371446 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371466 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371483 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371499 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371516 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371534 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371550 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371568 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371585 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371603 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371623 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371637 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371673 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371688 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371702 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371716 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371731 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371744 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371759 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371780 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371794 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371810 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371823 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371838 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371853 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371869 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371886 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371900 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371915 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371929 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371944 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371959 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371975 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371990 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372004 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372020 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372034 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372049 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372063 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372080 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372098 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372114 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372130 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372145 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372162 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372187 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372203 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372217 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372233 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372251 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372266 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372282 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372296 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372310 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372326 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372339 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372353 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372367 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372379 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372396 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372412 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372427 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372442 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372457 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372473 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372488 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372509 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372524 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372539 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372613 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372632 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372662 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372677 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372692 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372706 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372722 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372739 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372755 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372778 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372794 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372808 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372825 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372840 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372855 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372871 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372886 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372904 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372920 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372935 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372954 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372969 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372985 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.373000 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.373015 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.373031 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.373046 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.373062 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.373077 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.373094 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.373110 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.373126 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.373141 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.373156 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.373171 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.373188 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371070 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371280 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371412 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371496 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371801 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.371901 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372023 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372041 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372179 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.373261 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372258 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372261 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372398 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372429 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372442 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372542 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372806 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.373381 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372800 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.373117 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.373470 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.373167 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.372970 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.373420 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.373677 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.373686 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.373930 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.373973 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.373987 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.374064 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.374099 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.374069 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.374261 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.374881 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.374939 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.374981 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375082 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375126 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.373204 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375175 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375198 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375217 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375233 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375235 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375250 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375266 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375285 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375302 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375320 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375335 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375350 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375364 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375378 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375391 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375407 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375422 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375437 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375451 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375468 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375485 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375500 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375515 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375530 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375544 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375560 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375575 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375592 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375607 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375623 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375651 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375669 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375784 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375801 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375818 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375834 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375855 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375870 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375886 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375902 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375920 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375978 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375998 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376016 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376047 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376076 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376094 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376192 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376210 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376259 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376350 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376362 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376392 4559 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376402 4559 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376411 4559 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376421 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376429 4559 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376439 4559 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376448 4559 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376457 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376466 4559 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376573 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376584 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376593 4559 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376602 4559 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376611 4559 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376620 4559 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376628 4559 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376636 4559 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376662 4559 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376670 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376679 4559 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376688 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376696 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376706 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376714 4559 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376723 4559 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376731 4559 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376740 4559 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376747 4559 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376756 4559 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376766 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376783 4559 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376791 4559 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376800 4559 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376809 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376820 4559 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376828 4559 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376837 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376845 4559 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376677 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375262 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375258 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375291 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375412 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375514 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375528 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375562 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375587 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375723 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.382110 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.382230 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.375767 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376057 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376244 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376287 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376453 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376462 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.382291 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376587 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376703 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376877 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376919 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.376959 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.377244 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.377612 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.377880 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.377984 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:45:16.877968875 +0000 UTC m=+18.899954489 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.377998 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.378069 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.378088 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.378106 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.378125 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.378325 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.378335 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.378449 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.378483 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.378479 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.378493 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.378554 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.378609 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.378656 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.378664 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.378876 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.378924 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.378948 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.379005 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.379021 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.379169 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.379263 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.379316 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.379410 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.379483 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.379614 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.379765 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.379912 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.380027 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.380170 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.380223 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.380247 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.380291 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.380370 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.380379 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.380409 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.380473 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.380482 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.380516 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.380537 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.380854 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.380908 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.380983 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.381016 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.381162 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.381192 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.381201 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.381306 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.381528 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.381532 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.381550 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.381760 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.381854 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.381868 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.382625 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.381922 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.381926 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.381949 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.382004 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.382072 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.382349 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.382880 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.382944 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.382988 4559 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.383034 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:45:16.883021243 +0000 UTC m=+18.905006857 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.383066 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.383084 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.383115 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.383145 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.383251 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.383328 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.383335 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.383341 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.383363 4559 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.383422 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:45:16.88341035 +0000 UTC m=+18.905395964 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.383704 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.383997 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.384190 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.384618 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.384629 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.384851 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.384868 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.384885 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.385143 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.385173 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.385479 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.385499 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.385671 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.385692 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.385733 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.385835 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.385951 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.386553 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.386802 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.387782 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.389536 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.390788 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.390892 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.391165 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.391200 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.392758 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.392801 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.393007 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.393084 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.393119 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.393154 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.393346 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.393493 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.393604 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.393687 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.394049 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.394289 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.394302 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.394322 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.394353 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.394371 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.394440 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.394838 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.394849 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.395052 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.395355 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.395452 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.395538 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.395569 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.395926 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.396025 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.396122 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.396080 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.396399 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.396574 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.396583 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.406623 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.408289 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.417958 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.418456 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477550 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477621 4559 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477633 4559 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477656 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477665 4559 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477675 4559 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477683 4559 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477692 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477700 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477707 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477715 4559 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477722 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477729 4559 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477739 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477747 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477755 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477762 4559 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477776 4559 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477687 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477783 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477823 4559 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477833 4559 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477842 4559 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477851 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477859 4559 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477867 4559 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477875 4559 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477883 4559 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477890 4559 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477898 4559 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477906 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477914 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477922 4559 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477929 4559 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477937 4559 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477946 4559 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477954 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477962 4559 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477972 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477979 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477987 4559 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.477994 4559 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478001 4559 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478009 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478017 4559 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478024 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478032 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478039 4559 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478046 4559 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478054 4559 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478062 4559 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478071 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478079 4559 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478086 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478094 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478101 4559 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478109 4559 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478117 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478125 4559 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478133 4559 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478140 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478147 4559 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478155 4559 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478162 4559 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478169 4559 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478176 4559 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478183 4559 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478190 4559 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478197 4559 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478204 4559 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478213 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478220 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478227 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478233 4559 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478242 4559 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478249 4559 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478256 4559 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478264 4559 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478272 4559 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478280 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478292 4559 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478300 4559 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478308 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478316 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478323 4559 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478330 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478337 4559 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478346 4559 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478353 4559 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478361 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478368 4559 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478375 4559 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478382 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478390 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478398 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478405 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478412 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478421 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478428 4559 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478436 4559 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478443 4559 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478450 4559 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478458 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478465 4559 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478473 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478480 4559 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478487 4559 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478494 4559 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478501 4559 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478509 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478518 4559 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478526 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478536 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478545 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478552 4559 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478559 4559 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478566 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478574 4559 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478581 4559 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478589 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478597 4559 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478604 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478612 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478619 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478626 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478634 4559 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478657 4559 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478666 4559 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478673 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478681 4559 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478688 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478697 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478705 4559 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478712 4559 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478720 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478728 4559 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478735 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478742 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478751 4559 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478758 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478765 4559 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478784 4559 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478791 4559 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478799 4559 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478807 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478815 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478823 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478830 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478837 4559 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478844 4559 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478851 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478858 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478865 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478872 4559 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478879 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.478886 4559 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.555746 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.561415 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.565595 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 23 06:45:16 crc kubenswrapper[4559]: W1123 06:45:16.576492 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-a0ca8ec986bf62625df45b8f8955ab0c26bd2c2cb695b13a60ec0cf56544be27 WatchSource:0}: Error finding container a0ca8ec986bf62625df45b8f8955ab0c26bd2c2cb695b13a60ec0cf56544be27: Status 404 returned error can't find the container with id a0ca8ec986bf62625df45b8f8955ab0c26bd2c2cb695b13a60ec0cf56544be27 Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.780285 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.780405 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.780438 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.780455 4559 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.780498 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-23 06:45:17.780484815 +0000 UTC m=+19.802470439 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.881267 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.881340 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.881477 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.881504 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.881515 4559 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.881565 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-23 06:45:17.881551628 +0000 UTC m=+19.903537242 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.881714 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:45:17.881699044 +0000 UTC m=+19.903684668 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.982422 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:16 crc kubenswrapper[4559]: I1123 06:45:16.982457 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.982543 4559 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.982582 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:45:17.982571833 +0000 UTC m=+20.004557447 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.982781 4559 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:45:16 crc kubenswrapper[4559]: E1123 06:45:16.982878 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:45:17.982867877 +0000 UTC m=+20.004853481 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.272849 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:17 crc kubenswrapper[4559]: E1123 06:45:17.272958 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.336247 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.337413 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba"} Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.337801 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.338445 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"a0ca8ec986bf62625df45b8f8955ab0c26bd2c2cb695b13a60ec0cf56544be27"} Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.339259 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a"} Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.339300 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"cea58f9f53bb41889739c56ea8a004ee0f14be98328899e8396da68d80f9bda3"} Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.340447 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691"} Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.340477 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678"} Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.340487 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"828af8fe925ab386b42ff607a3a81060747143cfb7c0ff3518144282e964d61a"} Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.346892 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.357795 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.366353 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.374442 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.386006 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.393920 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.403536 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.413056 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.421023 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.429024 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.436938 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.446220 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.454978 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.462319 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:17Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.787373 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:17 crc kubenswrapper[4559]: E1123 06:45:17.787493 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:45:17 crc kubenswrapper[4559]: E1123 06:45:17.787508 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:45:17 crc kubenswrapper[4559]: E1123 06:45:17.787517 4559 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:45:17 crc kubenswrapper[4559]: E1123 06:45:17.787562 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-23 06:45:19.787550926 +0000 UTC m=+21.809536540 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.888406 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.888459 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:17 crc kubenswrapper[4559]: E1123 06:45:17.888554 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:45:19.888539443 +0000 UTC m=+21.910525067 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:45:17 crc kubenswrapper[4559]: E1123 06:45:17.888624 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:45:17 crc kubenswrapper[4559]: E1123 06:45:17.888638 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:45:17 crc kubenswrapper[4559]: E1123 06:45:17.888675 4559 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:45:17 crc kubenswrapper[4559]: E1123 06:45:17.888712 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-23 06:45:19.888704541 +0000 UTC m=+21.910690165 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.988913 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:17 crc kubenswrapper[4559]: I1123 06:45:17.988972 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:17 crc kubenswrapper[4559]: E1123 06:45:17.989022 4559 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:45:17 crc kubenswrapper[4559]: E1123 06:45:17.989050 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:45:19.989042613 +0000 UTC m=+22.011028217 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:45:17 crc kubenswrapper[4559]: E1123 06:45:17.989057 4559 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:45:17 crc kubenswrapper[4559]: E1123 06:45:17.989141 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:45:19.989122342 +0000 UTC m=+22.011107955 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.272615 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.272636 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:18 crc kubenswrapper[4559]: E1123 06:45:18.272805 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:45:18 crc kubenswrapper[4559]: E1123 06:45:18.272879 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.275410 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.276017 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.276967 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.277477 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.278315 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.278756 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.279261 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.280130 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.280655 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.281442 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.281596 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.281893 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.282788 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.283199 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.283621 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.284389 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.284861 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.285625 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.285978 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.286430 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.287264 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.287676 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.288473 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.288869 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.289260 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.289999 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.290545 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.291131 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.291730 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.292171 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.292695 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.293130 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.293550 4559 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.293663 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.294900 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.295325 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.295698 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.296665 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.297228 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.297705 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.298837 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.300808 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.301390 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.302162 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.302694 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.303559 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.304121 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.304878 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.305341 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.306146 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.306793 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.307408 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.307527 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.308048 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.308991 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.309436 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.309940 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.310675 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.315220 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.322903 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.330857 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.465974 4559 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.467102 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.467132 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.467142 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.467171 4559 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.471585 4559 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.471791 4559 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.472455 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.472481 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.472489 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.472502 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.472510 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:18Z","lastTransitionTime":"2025-11-23T06:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:18 crc kubenswrapper[4559]: E1123 06:45:18.484597 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.486896 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.487005 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.487081 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.487142 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.487191 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:18Z","lastTransitionTime":"2025-11-23T06:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:18 crc kubenswrapper[4559]: E1123 06:45:18.498365 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.500664 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.500699 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.500709 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.500722 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.500760 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:18Z","lastTransitionTime":"2025-11-23T06:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:18 crc kubenswrapper[4559]: E1123 06:45:18.509455 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.511635 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.511751 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.511825 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.511893 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.511943 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:18Z","lastTransitionTime":"2025-11-23T06:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:18 crc kubenswrapper[4559]: E1123 06:45:18.519458 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.521320 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.521353 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.521362 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.521389 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.521398 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:18Z","lastTransitionTime":"2025-11-23T06:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:18 crc kubenswrapper[4559]: E1123 06:45:18.528678 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:18 crc kubenswrapper[4559]: E1123 06:45:18.528786 4559 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.529556 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.529577 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.529585 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.529594 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.529600 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:18Z","lastTransitionTime":"2025-11-23T06:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.631581 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.631611 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.631620 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.631633 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.631658 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:18Z","lastTransitionTime":"2025-11-23T06:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.733835 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.733863 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.733872 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.733888 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.733896 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:18Z","lastTransitionTime":"2025-11-23T06:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.835911 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.835944 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.835969 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.835981 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.835988 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:18Z","lastTransitionTime":"2025-11-23T06:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.937972 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.938019 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.938029 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.938038 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:18 crc kubenswrapper[4559]: I1123 06:45:18.938047 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:18Z","lastTransitionTime":"2025-11-23T06:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.039383 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.039412 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.039420 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.039431 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.039439 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:19Z","lastTransitionTime":"2025-11-23T06:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.141212 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.141240 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.141249 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.141258 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.141266 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:19Z","lastTransitionTime":"2025-11-23T06:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.242544 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.242570 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.242596 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.242609 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.242619 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:19Z","lastTransitionTime":"2025-11-23T06:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.273378 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:19 crc kubenswrapper[4559]: E1123 06:45:19.273457 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.343657 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.343691 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.343699 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.343710 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.343718 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:19Z","lastTransitionTime":"2025-11-23T06:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.344629 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e"} Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.354826 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.378409 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.387320 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.396730 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.406056 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.414590 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.422442 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.445924 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.445949 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.445976 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.445988 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.445999 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:19Z","lastTransitionTime":"2025-11-23T06:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.547387 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.547420 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.547428 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.547460 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.547469 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:19Z","lastTransitionTime":"2025-11-23T06:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.649099 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.649145 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.649154 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.649167 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.649176 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:19Z","lastTransitionTime":"2025-11-23T06:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.751470 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.751511 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.751520 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.751534 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.751544 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:19Z","lastTransitionTime":"2025-11-23T06:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.803881 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:19 crc kubenswrapper[4559]: E1123 06:45:19.804581 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:45:19 crc kubenswrapper[4559]: E1123 06:45:19.804614 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:45:19 crc kubenswrapper[4559]: E1123 06:45:19.804629 4559 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:45:19 crc kubenswrapper[4559]: E1123 06:45:19.804701 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-23 06:45:23.804685366 +0000 UTC m=+25.826670980 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.853034 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.853072 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.853082 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.853094 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.853104 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:19Z","lastTransitionTime":"2025-11-23T06:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.905511 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.905594 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:19 crc kubenswrapper[4559]: E1123 06:45:19.905668 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:45:23.905650329 +0000 UTC m=+25.927635963 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:45:19 crc kubenswrapper[4559]: E1123 06:45:19.905701 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:45:19 crc kubenswrapper[4559]: E1123 06:45:19.905715 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:45:19 crc kubenswrapper[4559]: E1123 06:45:19.905735 4559 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:45:19 crc kubenswrapper[4559]: E1123 06:45:19.905770 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-23 06:45:23.905762619 +0000 UTC m=+25.927748232 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.955295 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.955328 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.955337 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.955350 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.955360 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:19Z","lastTransitionTime":"2025-11-23T06:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.958215 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-njqzh"] Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.958458 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-njqzh" Nov 23 06:45:19 crc kubenswrapper[4559]: W1123 06:45:19.959722 4559 reflector.go:561] object-"openshift-dns"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-dns": no relationship found between node 'crc' and this object Nov 23 06:45:19 crc kubenswrapper[4559]: E1123 06:45:19.959770 4559 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-dns\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:45:19 crc kubenswrapper[4559]: W1123 06:45:19.959735 4559 reflector.go:561] object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": failed to list *v1.Secret: secrets "node-resolver-dockercfg-kz9s7" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-dns": no relationship found between node 'crc' and this object Nov 23 06:45:19 crc kubenswrapper[4559]: E1123 06:45:19.959802 4559 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns\"/\"node-resolver-dockercfg-kz9s7\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"node-resolver-dockercfg-kz9s7\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-dns\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:45:19 crc kubenswrapper[4559]: W1123 06:45:19.960070 4559 reflector.go:561] object-"openshift-dns"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-dns": no relationship found between node 'crc' and this object Nov 23 06:45:19 crc kubenswrapper[4559]: E1123 06:45:19.960103 4559 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-dns\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.977119 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:19 crc kubenswrapper[4559]: I1123 06:45:19.993315 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:19Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.003484 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.005794 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.005817 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.005838 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvpfp\" (UniqueName: \"kubernetes.io/projected/16844d77-d909-4a27-bc7b-f99e97662ba3-kube-api-access-xvpfp\") pod \"node-resolver-njqzh\" (UID: \"16844d77-d909-4a27-bc7b-f99e97662ba3\") " pod="openshift-dns/node-resolver-njqzh" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.005853 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/16844d77-d909-4a27-bc7b-f99e97662ba3-hosts-file\") pod \"node-resolver-njqzh\" (UID: \"16844d77-d909-4a27-bc7b-f99e97662ba3\") " pod="openshift-dns/node-resolver-njqzh" Nov 23 06:45:20 crc kubenswrapper[4559]: E1123 06:45:20.005920 4559 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:45:20 crc kubenswrapper[4559]: E1123 06:45:20.005952 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:45:24.005942054 +0000 UTC m=+26.027927669 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:45:20 crc kubenswrapper[4559]: E1123 06:45:20.006141 4559 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:45:20 crc kubenswrapper[4559]: E1123 06:45:20.006179 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:45:24.006172164 +0000 UTC m=+26.028157778 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.013371 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.028392 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.036701 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.047253 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.057007 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.057035 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.057042 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.057056 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.057064 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:20Z","lastTransitionTime":"2025-11-23T06:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.058773 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.107127 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvpfp\" (UniqueName: \"kubernetes.io/projected/16844d77-d909-4a27-bc7b-f99e97662ba3-kube-api-access-xvpfp\") pod \"node-resolver-njqzh\" (UID: \"16844d77-d909-4a27-bc7b-f99e97662ba3\") " pod="openshift-dns/node-resolver-njqzh" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.107255 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/16844d77-d909-4a27-bc7b-f99e97662ba3-hosts-file\") pod \"node-resolver-njqzh\" (UID: \"16844d77-d909-4a27-bc7b-f99e97662ba3\") " pod="openshift-dns/node-resolver-njqzh" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.107363 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/16844d77-d909-4a27-bc7b-f99e97662ba3-hosts-file\") pod \"node-resolver-njqzh\" (UID: \"16844d77-d909-4a27-bc7b-f99e97662ba3\") " pod="openshift-dns/node-resolver-njqzh" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.159069 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.159100 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.159108 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.159121 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.159129 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:20Z","lastTransitionTime":"2025-11-23T06:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.260955 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.260988 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.260996 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.261008 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.261017 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:20Z","lastTransitionTime":"2025-11-23T06:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.273472 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:20 crc kubenswrapper[4559]: E1123 06:45:20.273580 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.273477 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:20 crc kubenswrapper[4559]: E1123 06:45:20.273685 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.362742 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.362779 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.362788 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.362801 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.362810 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:20Z","lastTransitionTime":"2025-11-23T06:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.465196 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.465235 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.465243 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.465256 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.465265 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:20Z","lastTransitionTime":"2025-11-23T06:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.536995 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.539543 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.544894 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.551860 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.552705 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.566939 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.566967 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.566975 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.566987 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.566996 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:20Z","lastTransitionTime":"2025-11-23T06:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.567237 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.578999 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.588789 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.599285 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.611970 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.620716 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.628668 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.636503 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.643843 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.651626 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.662138 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.668546 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.668575 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.668584 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.668599 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.668607 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:20Z","lastTransitionTime":"2025-11-23T06:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.670884 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.680086 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.688183 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.698473 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.703455 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-s7zkt"] Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.703755 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-qn4h8"] Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.703929 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.704070 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-rlxnw"] Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.704262 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.704980 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.705727 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jt56f"] Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.706244 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.706359 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.706688 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 23 06:45:20 crc kubenswrapper[4559]: W1123 06:45:20.707139 4559 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovnkube-config": failed to list *v1.ConfigMap: configmaps "ovnkube-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.707174 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.707206 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.707145 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 23 06:45:20 crc kubenswrapper[4559]: E1123 06:45:20.707174 4559 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovnkube-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"ovnkube-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.707333 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.707377 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.707416 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.707526 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.707660 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.708582 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.708659 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.708866 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.709498 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.711084 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.711174 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.711196 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.711251 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.719931 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.729202 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.737278 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.745544 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.753511 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.765168 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.769972 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.769997 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.770006 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.770018 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.770026 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:20Z","lastTransitionTime":"2025-11-23T06:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.784864 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.806930 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.811758 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-multus-cni-dir\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.811786 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-run-ovn\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.811810 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e9164873-0a60-498c-9b37-cd73817b066a-cnibin\") pod \"multus-additional-cni-plugins-rlxnw\" (UID: \"e9164873-0a60-498c-9b37-cd73817b066a\") " pod="openshift-multus/multus-additional-cni-plugins-rlxnw" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.811824 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/18091d80-4837-4894-8583-9a2bd30e10b9-multus-daemon-config\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.811839 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtqzh\" (UniqueName: \"kubernetes.io/projected/18091d80-4837-4894-8583-9a2bd30e10b9-kube-api-access-wtqzh\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.811854 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-etc-openvswitch\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.811866 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-node-log\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.811880 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b12e257e-7887-4795-9221-4db8fd5856dd-env-overrides\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.811909 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/18091d80-4837-4894-8583-9a2bd30e10b9-cni-binary-copy\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.811923 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-multus-socket-dir-parent\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.811936 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-run-openvswitch\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.811952 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/4731beee-0cac-4189-8a70-743b0b709095-rootfs\") pod \"machine-config-daemon-s7zkt\" (UID: \"4731beee-0cac-4189-8a70-743b0b709095\") " pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.811966 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd57x\" (UniqueName: \"kubernetes.io/projected/4731beee-0cac-4189-8a70-743b0b709095-kube-api-access-bd57x\") pod \"machine-config-daemon-s7zkt\" (UID: \"4731beee-0cac-4189-8a70-743b0b709095\") " pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.811980 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-host-run-multus-certs\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.811995 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e9164873-0a60-498c-9b37-cd73817b066a-system-cni-dir\") pod \"multus-additional-cni-plugins-rlxnw\" (UID: \"e9164873-0a60-498c-9b37-cd73817b066a\") " pod="openshift-multus/multus-additional-cni-plugins-rlxnw" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812007 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-host-var-lib-kubelet\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812020 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-systemd-units\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812036 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4731beee-0cac-4189-8a70-743b0b709095-mcd-auth-proxy-config\") pod \"machine-config-daemon-s7zkt\" (UID: \"4731beee-0cac-4189-8a70-743b0b709095\") " pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812049 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-run-systemd\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812064 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812078 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcdbt\" (UniqueName: \"kubernetes.io/projected/b12e257e-7887-4795-9221-4db8fd5856dd-kube-api-access-jcdbt\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812091 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e9164873-0a60-498c-9b37-cd73817b066a-cni-binary-copy\") pod \"multus-additional-cni-plugins-rlxnw\" (UID: \"e9164873-0a60-498c-9b37-cd73817b066a\") " pod="openshift-multus/multus-additional-cni-plugins-rlxnw" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812103 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e9164873-0a60-498c-9b37-cd73817b066a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rlxnw\" (UID: \"e9164873-0a60-498c-9b37-cd73817b066a\") " pod="openshift-multus/multus-additional-cni-plugins-rlxnw" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812116 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4731beee-0cac-4189-8a70-743b0b709095-proxy-tls\") pod \"machine-config-daemon-s7zkt\" (UID: \"4731beee-0cac-4189-8a70-743b0b709095\") " pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812135 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-host-run-k8s-cni-cncf-io\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812149 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-cni-bin\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812169 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-multus-conf-dir\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812181 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-run-ovn-kubernetes\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812193 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-cni-netd\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812212 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e9164873-0a60-498c-9b37-cd73817b066a-os-release\") pod \"multus-additional-cni-plugins-rlxnw\" (UID: \"e9164873-0a60-498c-9b37-cd73817b066a\") " pod="openshift-multus/multus-additional-cni-plugins-rlxnw" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812226 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-log-socket\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812239 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-os-release\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812252 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-host-var-lib-cni-multus\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812266 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-hostroot\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812277 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-etc-kubernetes\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812291 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-kubelet\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812302 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b12e257e-7887-4795-9221-4db8fd5856dd-ovnkube-config\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812321 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-system-cni-dir\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812332 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-cnibin\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812344 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b12e257e-7887-4795-9221-4db8fd5856dd-ovn-node-metrics-cert\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812357 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b12e257e-7887-4795-9221-4db8fd5856dd-ovnkube-script-lib\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812376 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e9164873-0a60-498c-9b37-cd73817b066a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rlxnw\" (UID: \"e9164873-0a60-498c-9b37-cd73817b066a\") " pod="openshift-multus/multus-additional-cni-plugins-rlxnw" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812388 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7s7s\" (UniqueName: \"kubernetes.io/projected/e9164873-0a60-498c-9b37-cd73817b066a-kube-api-access-n7s7s\") pod \"multus-additional-cni-plugins-rlxnw\" (UID: \"e9164873-0a60-498c-9b37-cd73817b066a\") " pod="openshift-multus/multus-additional-cni-plugins-rlxnw" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812402 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-host-run-netns\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812414 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-host-var-lib-cni-bin\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812427 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-slash\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812439 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-run-netns\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.812454 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-var-lib-openvswitch\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.819249 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.824793 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.831053 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.841631 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.855114 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.862757 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.872084 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.872108 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.872116 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.872128 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.872137 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:20Z","lastTransitionTime":"2025-11-23T06:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.876823 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.884931 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.892040 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.900889 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.908896 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.913691 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b12e257e-7887-4795-9221-4db8fd5856dd-env-overrides\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.913731 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e9164873-0a60-498c-9b37-cd73817b066a-cnibin\") pod \"multus-additional-cni-plugins-rlxnw\" (UID: \"e9164873-0a60-498c-9b37-cd73817b066a\") " pod="openshift-multus/multus-additional-cni-plugins-rlxnw" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.913747 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/18091d80-4837-4894-8583-9a2bd30e10b9-multus-daemon-config\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.913762 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtqzh\" (UniqueName: \"kubernetes.io/projected/18091d80-4837-4894-8583-9a2bd30e10b9-kube-api-access-wtqzh\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.913776 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-etc-openvswitch\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.913788 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-node-log\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.913803 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/18091d80-4837-4894-8583-9a2bd30e10b9-cni-binary-copy\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.913818 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-multus-socket-dir-parent\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.913830 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-run-openvswitch\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.913846 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/4731beee-0cac-4189-8a70-743b0b709095-rootfs\") pod \"machine-config-daemon-s7zkt\" (UID: \"4731beee-0cac-4189-8a70-743b0b709095\") " pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.913862 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bd57x\" (UniqueName: \"kubernetes.io/projected/4731beee-0cac-4189-8a70-743b0b709095-kube-api-access-bd57x\") pod \"machine-config-daemon-s7zkt\" (UID: \"4731beee-0cac-4189-8a70-743b0b709095\") " pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.913876 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-host-run-multus-certs\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.913888 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-systemd-units\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.913892 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-node-log\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.913917 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-run-openvswitch\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.913943 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-systemd-units\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.913903 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e9164873-0a60-498c-9b37-cd73817b066a-cnibin\") pod \"multus-additional-cni-plugins-rlxnw\" (UID: \"e9164873-0a60-498c-9b37-cd73817b066a\") " pod="openshift-multus/multus-additional-cni-plugins-rlxnw" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.913925 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e9164873-0a60-498c-9b37-cd73817b066a-system-cni-dir\") pod \"multus-additional-cni-plugins-rlxnw\" (UID: \"e9164873-0a60-498c-9b37-cd73817b066a\") " pod="openshift-multus/multus-additional-cni-plugins-rlxnw" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.913957 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-host-run-multus-certs\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.913905 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e9164873-0a60-498c-9b37-cd73817b066a-system-cni-dir\") pod \"multus-additional-cni-plugins-rlxnw\" (UID: \"e9164873-0a60-498c-9b37-cd73817b066a\") " pod="openshift-multus/multus-additional-cni-plugins-rlxnw" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.913985 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-host-var-lib-kubelet\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.913992 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/4731beee-0cac-4189-8a70-743b0b709095-rootfs\") pod \"machine-config-daemon-s7zkt\" (UID: \"4731beee-0cac-4189-8a70-743b0b709095\") " pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914000 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914020 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcdbt\" (UniqueName: \"kubernetes.io/projected/b12e257e-7887-4795-9221-4db8fd5856dd-kube-api-access-jcdbt\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914031 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-multus-socket-dir-parent\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914037 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4731beee-0cac-4189-8a70-743b0b709095-mcd-auth-proxy-config\") pod \"machine-config-daemon-s7zkt\" (UID: \"4731beee-0cac-4189-8a70-743b0b709095\") " pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914037 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-host-var-lib-kubelet\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914051 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-run-systemd\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914060 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914068 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e9164873-0a60-498c-9b37-cd73817b066a-cni-binary-copy\") pod \"multus-additional-cni-plugins-rlxnw\" (UID: \"e9164873-0a60-498c-9b37-cd73817b066a\") " pod="openshift-multus/multus-additional-cni-plugins-rlxnw" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914086 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e9164873-0a60-498c-9b37-cd73817b066a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rlxnw\" (UID: \"e9164873-0a60-498c-9b37-cd73817b066a\") " pod="openshift-multus/multus-additional-cni-plugins-rlxnw" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914101 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4731beee-0cac-4189-8a70-743b0b709095-proxy-tls\") pod \"machine-config-daemon-s7zkt\" (UID: \"4731beee-0cac-4189-8a70-743b0b709095\") " pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914116 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-host-run-k8s-cni-cncf-io\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914129 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-cni-bin\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914142 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-cni-netd\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914164 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-multus-conf-dir\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914177 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-run-ovn-kubernetes\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914195 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e9164873-0a60-498c-9b37-cd73817b066a-os-release\") pod \"multus-additional-cni-plugins-rlxnw\" (UID: \"e9164873-0a60-498c-9b37-cd73817b066a\") " pod="openshift-multus/multus-additional-cni-plugins-rlxnw" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914207 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-log-socket\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914207 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-host-run-k8s-cni-cncf-io\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914219 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b12e257e-7887-4795-9221-4db8fd5856dd-ovnkube-config\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914222 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b12e257e-7887-4795-9221-4db8fd5856dd-env-overrides\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914234 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-os-release\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914248 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-host-var-lib-cni-multus\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914262 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-hostroot\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914267 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-run-ovn-kubernetes\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914275 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-etc-kubernetes\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914290 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-kubelet\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914296 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-cni-bin\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914311 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-system-cni-dir\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914318 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-cni-netd\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914324 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-cnibin\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914340 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-multus-conf-dir\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914340 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b12e257e-7887-4795-9221-4db8fd5856dd-ovn-node-metrics-cert\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914371 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b12e257e-7887-4795-9221-4db8fd5856dd-ovnkube-script-lib\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914375 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/18091d80-4837-4894-8583-9a2bd30e10b9-multus-daemon-config\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914397 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e9164873-0a60-498c-9b37-cd73817b066a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rlxnw\" (UID: \"e9164873-0a60-498c-9b37-cd73817b066a\") " pod="openshift-multus/multus-additional-cni-plugins-rlxnw" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914418 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7s7s\" (UniqueName: \"kubernetes.io/projected/e9164873-0a60-498c-9b37-cd73817b066a-kube-api-access-n7s7s\") pod \"multus-additional-cni-plugins-rlxnw\" (UID: \"e9164873-0a60-498c-9b37-cd73817b066a\") " pod="openshift-multus/multus-additional-cni-plugins-rlxnw" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914432 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-host-run-netns\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914445 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-host-var-lib-cni-bin\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914458 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-slash\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914472 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-run-netns\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914471 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/18091d80-4837-4894-8583-9a2bd30e10b9-cni-binary-copy\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914493 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-var-lib-openvswitch\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914086 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-run-systemd\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914511 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-multus-cni-dir\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914523 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-run-ovn\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914530 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-host-var-lib-cni-multus\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914565 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-log-socket\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914602 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-os-release\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914610 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e9164873-0a60-498c-9b37-cd73817b066a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rlxnw\" (UID: \"e9164873-0a60-498c-9b37-cd73817b066a\") " pod="openshift-multus/multus-additional-cni-plugins-rlxnw" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914657 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-run-netns\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914695 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4731beee-0cac-4189-8a70-743b0b709095-mcd-auth-proxy-config\") pod \"machine-config-daemon-s7zkt\" (UID: \"4731beee-0cac-4189-8a70-743b0b709095\") " pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914729 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e9164873-0a60-498c-9b37-cd73817b066a-cni-binary-copy\") pod \"multus-additional-cni-plugins-rlxnw\" (UID: \"e9164873-0a60-498c-9b37-cd73817b066a\") " pod="openshift-multus/multus-additional-cni-plugins-rlxnw" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914747 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-host-var-lib-cni-bin\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914768 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-host-run-netns\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914786 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-run-ovn\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914799 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-slash\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914807 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-var-lib-openvswitch\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914849 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e9164873-0a60-498c-9b37-cd73817b066a-os-release\") pod \"multus-additional-cni-plugins-rlxnw\" (UID: \"e9164873-0a60-498c-9b37-cd73817b066a\") " pod="openshift-multus/multus-additional-cni-plugins-rlxnw" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914853 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-multus-cni-dir\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914868 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-system-cni-dir\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914869 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-kubelet\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914882 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-hostroot\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914882 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-etc-kubernetes\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914937 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b12e257e-7887-4795-9221-4db8fd5856dd-ovnkube-script-lib\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.914923 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/18091d80-4837-4894-8583-9a2bd30e10b9-cnibin\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.915013 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e9164873-0a60-498c-9b37-cd73817b066a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rlxnw\" (UID: \"e9164873-0a60-498c-9b37-cd73817b066a\") " pod="openshift-multus/multus-additional-cni-plugins-rlxnw" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.915021 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-etc-openvswitch\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.917699 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.918838 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4731beee-0cac-4189-8a70-743b0b709095-proxy-tls\") pod \"machine-config-daemon-s7zkt\" (UID: \"4731beee-0cac-4189-8a70-743b0b709095\") " pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.925480 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b12e257e-7887-4795-9221-4db8fd5856dd-ovn-node-metrics-cert\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.926035 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd57x\" (UniqueName: \"kubernetes.io/projected/4731beee-0cac-4189-8a70-743b0b709095-kube-api-access-bd57x\") pod \"machine-config-daemon-s7zkt\" (UID: \"4731beee-0cac-4189-8a70-743b0b709095\") " pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.926789 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7s7s\" (UniqueName: \"kubernetes.io/projected/e9164873-0a60-498c-9b37-cd73817b066a-kube-api-access-n7s7s\") pod \"multus-additional-cni-plugins-rlxnw\" (UID: \"e9164873-0a60-498c-9b37-cd73817b066a\") " pod="openshift-multus/multus-additional-cni-plugins-rlxnw" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.927065 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.927351 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcdbt\" (UniqueName: \"kubernetes.io/projected/b12e257e-7887-4795-9221-4db8fd5856dd-kube-api-access-jcdbt\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.931900 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtqzh\" (UniqueName: \"kubernetes.io/projected/18091d80-4837-4894-8583-9a2bd30e10b9-kube-api-access-wtqzh\") pod \"multus-qn4h8\" (UID: \"18091d80-4837-4894-8583-9a2bd30e10b9\") " pod="openshift-multus/multus-qn4h8" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.936094 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.943907 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.951902 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.959433 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.974510 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.974534 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.974542 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.974553 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:20 crc kubenswrapper[4559]: I1123 06:45:20.974561 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:20Z","lastTransitionTime":"2025-11-23T06:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.014174 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-qn4h8" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.019017 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.024368 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" Nov 23 06:45:21 crc kubenswrapper[4559]: W1123 06:45:21.024548 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18091d80_4837_4894_8583_9a2bd30e10b9.slice/crio-b4a9f1d1ffac2efad53fef8206e6ff5ba67d517807c230241ce0e5d028cb5199 WatchSource:0}: Error finding container b4a9f1d1ffac2efad53fef8206e6ff5ba67d517807c230241ce0e5d028cb5199: Status 404 returned error can't find the container with id b4a9f1d1ffac2efad53fef8206e6ff5ba67d517807c230241ce0e5d028cb5199 Nov 23 06:45:21 crc kubenswrapper[4559]: W1123 06:45:21.030893 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4731beee_0cac_4189_8a70_743b0b709095.slice/crio-c294d916800c298bba5419cda06bd5111d3c6ef3a4bfc86b4913d9cb4ca96fb9 WatchSource:0}: Error finding container c294d916800c298bba5419cda06bd5111d3c6ef3a4bfc86b4913d9cb4ca96fb9: Status 404 returned error can't find the container with id c294d916800c298bba5419cda06bd5111d3c6ef3a4bfc86b4913d9cb4ca96fb9 Nov 23 06:45:21 crc kubenswrapper[4559]: W1123 06:45:21.036947 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9164873_0a60_498c_9b37_cd73817b066a.slice/crio-19a498e3572fdd6c5c88320e1994cb1a52a72725aa0d450c13f978338468b86d WatchSource:0}: Error finding container 19a498e3572fdd6c5c88320e1994cb1a52a72725aa0d450c13f978338468b86d: Status 404 returned error can't find the container with id 19a498e3572fdd6c5c88320e1994cb1a52a72725aa0d450c13f978338468b86d Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.075724 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.075758 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.075767 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.075780 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.075789 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:21Z","lastTransitionTime":"2025-11-23T06:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.167923 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.178684 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.178721 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.178729 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.178751 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.178759 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:21Z","lastTransitionTime":"2025-11-23T06:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.222792 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.230894 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvpfp\" (UniqueName: \"kubernetes.io/projected/16844d77-d909-4a27-bc7b-f99e97662ba3-kube-api-access-xvpfp\") pod \"node-resolver-njqzh\" (UID: \"16844d77-d909-4a27-bc7b-f99e97662ba3\") " pod="openshift-dns/node-resolver-njqzh" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.272577 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:21 crc kubenswrapper[4559]: E1123 06:45:21.272907 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.281234 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.281265 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.281274 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.281287 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.281296 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:21Z","lastTransitionTime":"2025-11-23T06:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.349157 4559 generic.go:334] "Generic (PLEG): container finished" podID="e9164873-0a60-498c-9b37-cd73817b066a" containerID="3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0" exitCode=0 Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.349208 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" event={"ID":"e9164873-0a60-498c-9b37-cd73817b066a","Type":"ContainerDied","Data":"3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0"} Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.349229 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" event={"ID":"e9164873-0a60-498c-9b37-cd73817b066a","Type":"ContainerStarted","Data":"19a498e3572fdd6c5c88320e1994cb1a52a72725aa0d450c13f978338468b86d"} Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.351392 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerStarted","Data":"329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a"} Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.351436 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerStarted","Data":"053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43"} Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.351446 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerStarted","Data":"c294d916800c298bba5419cda06bd5111d3c6ef3a4bfc86b4913d9cb4ca96fb9"} Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.352730 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qn4h8" event={"ID":"18091d80-4837-4894-8583-9a2bd30e10b9","Type":"ContainerStarted","Data":"aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64"} Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.352778 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qn4h8" event={"ID":"18091d80-4837-4894-8583-9a2bd30e10b9","Type":"ContainerStarted","Data":"b4a9f1d1ffac2efad53fef8206e6ff5ba67d517807c230241ce0e5d028cb5199"} Nov 23 06:45:21 crc kubenswrapper[4559]: E1123 06:45:21.356793 4559 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.358602 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.369886 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.379951 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.385787 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.385818 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.385828 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.385839 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.385848 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:21Z","lastTransitionTime":"2025-11-23T06:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.389551 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.399455 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.408662 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.416590 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.427088 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.438088 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.445876 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.458016 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.464739 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.466989 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-njqzh" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.473407 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:21 crc kubenswrapper[4559]: W1123 06:45:21.476935 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16844d77_d909_4a27_bc7b_f99e97662ba3.slice/crio-b3fc5fa771cb4ed03b2d991a7cb8cd01d11040255129c1ee5bd882b2dd6f0e70 WatchSource:0}: Error finding container b3fc5fa771cb4ed03b2d991a7cb8cd01d11040255129c1ee5bd882b2dd6f0e70: Status 404 returned error can't find the container with id b3fc5fa771cb4ed03b2d991a7cb8cd01d11040255129c1ee5bd882b2dd6f0e70 Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.487808 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.487840 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.487848 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.487860 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.487868 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:21Z","lastTransitionTime":"2025-11-23T06:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.488607 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.498897 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.510535 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.525303 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.534443 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.543472 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.558530 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.585306 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.589487 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.589513 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.589521 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.589532 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.589540 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:21Z","lastTransitionTime":"2025-11-23T06:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.623261 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.667287 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.678369 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.686103 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b12e257e-7887-4795-9221-4db8fd5856dd-ovnkube-config\") pod \"ovnkube-node-jt56f\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.691946 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.691981 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.691990 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.692003 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.692011 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:21Z","lastTransitionTime":"2025-11-23T06:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.723825 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.764662 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.793843 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.793871 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.793880 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.793893 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.793901 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:21Z","lastTransitionTime":"2025-11-23T06:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.804256 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:21Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.895723 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.895760 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.895770 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.895782 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.895790 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:21Z","lastTransitionTime":"2025-11-23T06:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.930132 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.997482 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.997516 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.997525 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.997539 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:21 crc kubenswrapper[4559]: I1123 06:45:21.997549 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:21Z","lastTransitionTime":"2025-11-23T06:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.099988 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.100174 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.100185 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.100199 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.100207 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:22Z","lastTransitionTime":"2025-11-23T06:45:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.202051 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.202084 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.202093 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.202105 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.202114 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:22Z","lastTransitionTime":"2025-11-23T06:45:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.272848 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.272951 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:22 crc kubenswrapper[4559]: E1123 06:45:22.272985 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:45:22 crc kubenswrapper[4559]: E1123 06:45:22.273058 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.303743 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.303772 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.303781 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.303793 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.303802 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:22Z","lastTransitionTime":"2025-11-23T06:45:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.356455 4559 generic.go:334] "Generic (PLEG): container finished" podID="e9164873-0a60-498c-9b37-cd73817b066a" containerID="1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb" exitCode=0 Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.356525 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" event={"ID":"e9164873-0a60-498c-9b37-cd73817b066a","Type":"ContainerDied","Data":"1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb"} Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.357678 4559 generic.go:334] "Generic (PLEG): container finished" podID="b12e257e-7887-4795-9221-4db8fd5856dd" containerID="5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d" exitCode=0 Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.357744 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" event={"ID":"b12e257e-7887-4795-9221-4db8fd5856dd","Type":"ContainerDied","Data":"5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d"} Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.357774 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" event={"ID":"b12e257e-7887-4795-9221-4db8fd5856dd","Type":"ContainerStarted","Data":"f87f51d90c99a905784a62114224a7ff2e76c7d1929991399bda41703cc39d45"} Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.358899 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-njqzh" event={"ID":"16844d77-d909-4a27-bc7b-f99e97662ba3","Type":"ContainerStarted","Data":"bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4"} Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.358925 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-njqzh" event={"ID":"16844d77-d909-4a27-bc7b-f99e97662ba3","Type":"ContainerStarted","Data":"b3fc5fa771cb4ed03b2d991a7cb8cd01d11040255129c1ee5bd882b2dd6f0e70"} Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.366249 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:22Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.373185 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:22Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.385076 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:22Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.393475 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:22Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.402355 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:22Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.404944 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.404972 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.404980 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.404992 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.405001 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:22Z","lastTransitionTime":"2025-11-23T06:45:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.411399 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:22Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.419485 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:22Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.427732 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:22Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.435611 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:22Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.445402 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:22Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.456291 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:22Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.463975 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:22Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.476537 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:22Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.485114 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:22Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.493378 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:22Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.502702 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:22Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.508189 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.508214 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.508222 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.508233 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.508241 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:22Z","lastTransitionTime":"2025-11-23T06:45:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.510430 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:22Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.524921 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:22Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.565941 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:22Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.609851 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.609890 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.609900 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.609915 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.609924 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:22Z","lastTransitionTime":"2025-11-23T06:45:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.623747 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:22Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.649041 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:22Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.687716 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:22Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.711340 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.711378 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.711389 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.711402 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.711410 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:22Z","lastTransitionTime":"2025-11-23T06:45:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.724553 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:22Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.763436 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:22Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.805241 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:22Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.813633 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.813684 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.813701 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.813713 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.813722 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:22Z","lastTransitionTime":"2025-11-23T06:45:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.844755 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:22Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.885977 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-w5tlv"] Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.886296 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-w5tlv" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.888331 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.898505 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.915732 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.915777 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.915787 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.915801 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.915810 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:22Z","lastTransitionTime":"2025-11-23T06:45:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.918821 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.930466 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/cdf91a44-8c01-4104-8d6f-e76b8c321166-serviceca\") pod \"node-ca-w5tlv\" (UID: \"cdf91a44-8c01-4104-8d6f-e76b8c321166\") " pod="openshift-image-registry/node-ca-w5tlv" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.930500 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cdf91a44-8c01-4104-8d6f-e76b8c321166-host\") pod \"node-ca-w5tlv\" (UID: \"cdf91a44-8c01-4104-8d6f-e76b8c321166\") " pod="openshift-image-registry/node-ca-w5tlv" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.930545 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7g4kl\" (UniqueName: \"kubernetes.io/projected/cdf91a44-8c01-4104-8d6f-e76b8c321166-kube-api-access-7g4kl\") pod \"node-ca-w5tlv\" (UID: \"cdf91a44-8c01-4104-8d6f-e76b8c321166\") " pod="openshift-image-registry/node-ca-w5tlv" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.938552 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 23 06:45:22 crc kubenswrapper[4559]: I1123 06:45:22.964723 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:22Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.006622 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:23Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.017505 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.017537 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.017557 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.017572 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.017583 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:23Z","lastTransitionTime":"2025-11-23T06:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.031340 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/cdf91a44-8c01-4104-8d6f-e76b8c321166-serviceca\") pod \"node-ca-w5tlv\" (UID: \"cdf91a44-8c01-4104-8d6f-e76b8c321166\") " pod="openshift-image-registry/node-ca-w5tlv" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.031380 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cdf91a44-8c01-4104-8d6f-e76b8c321166-host\") pod \"node-ca-w5tlv\" (UID: \"cdf91a44-8c01-4104-8d6f-e76b8c321166\") " pod="openshift-image-registry/node-ca-w5tlv" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.031422 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7g4kl\" (UniqueName: \"kubernetes.io/projected/cdf91a44-8c01-4104-8d6f-e76b8c321166-kube-api-access-7g4kl\") pod \"node-ca-w5tlv\" (UID: \"cdf91a44-8c01-4104-8d6f-e76b8c321166\") " pod="openshift-image-registry/node-ca-w5tlv" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.031570 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cdf91a44-8c01-4104-8d6f-e76b8c321166-host\") pod \"node-ca-w5tlv\" (UID: \"cdf91a44-8c01-4104-8d6f-e76b8c321166\") " pod="openshift-image-registry/node-ca-w5tlv" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.032358 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/cdf91a44-8c01-4104-8d6f-e76b8c321166-serviceca\") pod \"node-ca-w5tlv\" (UID: \"cdf91a44-8c01-4104-8d6f-e76b8c321166\") " pod="openshift-image-registry/node-ca-w5tlv" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.047066 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:23Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.072494 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7g4kl\" (UniqueName: \"kubernetes.io/projected/cdf91a44-8c01-4104-8d6f-e76b8c321166-kube-api-access-7g4kl\") pod \"node-ca-w5tlv\" (UID: \"cdf91a44-8c01-4104-8d6f-e76b8c321166\") " pod="openshift-image-registry/node-ca-w5tlv" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.106753 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:23Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.119287 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.119390 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.119454 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.119522 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.119580 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:23Z","lastTransitionTime":"2025-11-23T06:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.144790 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:23Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.185827 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:23Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.196196 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-w5tlv" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.222973 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.223006 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.223015 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.223028 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.223039 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:23Z","lastTransitionTime":"2025-11-23T06:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.225440 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:23Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.265917 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:23Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.273025 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:23 crc kubenswrapper[4559]: E1123 06:45:23.273107 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.310264 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:23Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.325361 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.325433 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.325449 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.325474 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.325488 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:23Z","lastTransitionTime":"2025-11-23T06:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.346448 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:23Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.364212 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" event={"ID":"b12e257e-7887-4795-9221-4db8fd5856dd","Type":"ContainerStarted","Data":"1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb"} Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.364240 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" event={"ID":"b12e257e-7887-4795-9221-4db8fd5856dd","Type":"ContainerStarted","Data":"664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3"} Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.364251 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" event={"ID":"b12e257e-7887-4795-9221-4db8fd5856dd","Type":"ContainerStarted","Data":"36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d"} Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.364259 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" event={"ID":"b12e257e-7887-4795-9221-4db8fd5856dd","Type":"ContainerStarted","Data":"ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77"} Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.364267 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" event={"ID":"b12e257e-7887-4795-9221-4db8fd5856dd","Type":"ContainerStarted","Data":"651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565"} Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.364275 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" event={"ID":"b12e257e-7887-4795-9221-4db8fd5856dd","Type":"ContainerStarted","Data":"96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb"} Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.365710 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-w5tlv" event={"ID":"cdf91a44-8c01-4104-8d6f-e76b8c321166","Type":"ContainerStarted","Data":"0b1b0de90eeea4fce9fc5783f6758c7563102beefca727f743b8db0842eaaca5"} Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.368299 4559 generic.go:334] "Generic (PLEG): container finished" podID="e9164873-0a60-498c-9b37-cd73817b066a" containerID="0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125" exitCode=0 Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.368330 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" event={"ID":"e9164873-0a60-498c-9b37-cd73817b066a","Type":"ContainerDied","Data":"0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125"} Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.384457 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:23Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.425576 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:23Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.427806 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.427837 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.427848 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.427862 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.427872 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:23Z","lastTransitionTime":"2025-11-23T06:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.465436 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:23Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.504404 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:23Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.530778 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.530832 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.530850 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.530877 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.530893 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:23Z","lastTransitionTime":"2025-11-23T06:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.544068 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:23Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.585379 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:23Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.625985 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:23Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.633665 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.633710 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.633722 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.633741 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.633752 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:23Z","lastTransitionTime":"2025-11-23T06:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.664048 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:23Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.703723 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:23Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.736006 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.736045 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.736053 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.736068 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.736079 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:23Z","lastTransitionTime":"2025-11-23T06:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.744244 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:23Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.785415 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:23Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.824885 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:23Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.837920 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:23 crc kubenswrapper[4559]: E1123 06:45:23.838202 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:45:23 crc kubenswrapper[4559]: E1123 06:45:23.838239 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:45:23 crc kubenswrapper[4559]: E1123 06:45:23.838277 4559 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.838454 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:23 crc kubenswrapper[4559]: E1123 06:45:23.838476 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-23 06:45:31.838455354 +0000 UTC m=+33.860440968 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.838482 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.838604 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.838627 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.838638 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:23Z","lastTransitionTime":"2025-11-23T06:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.864475 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:23Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.907629 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:23Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.938756 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:45:23 crc kubenswrapper[4559]: E1123 06:45:23.938916 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:45:31.938898052 +0000 UTC m=+33.960883666 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.938947 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:23 crc kubenswrapper[4559]: E1123 06:45:23.939076 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:45:23 crc kubenswrapper[4559]: E1123 06:45:23.939103 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:45:23 crc kubenswrapper[4559]: E1123 06:45:23.939115 4559 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:45:23 crc kubenswrapper[4559]: E1123 06:45:23.939164 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-23 06:45:31.939149081 +0000 UTC m=+33.961134695 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.940052 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.940075 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.940085 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.940103 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.940112 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:23Z","lastTransitionTime":"2025-11-23T06:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.943703 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:23Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:23 crc kubenswrapper[4559]: I1123 06:45:23.983237 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:23Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.023659 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:24Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.039435 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.039474 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:24 crc kubenswrapper[4559]: E1123 06:45:24.039561 4559 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:45:24 crc kubenswrapper[4559]: E1123 06:45:24.039594 4559 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:45:24 crc kubenswrapper[4559]: E1123 06:45:24.039617 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:45:32.039602427 +0000 UTC m=+34.061588041 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:45:24 crc kubenswrapper[4559]: E1123 06:45:24.039632 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:45:32.039623627 +0000 UTC m=+34.061609241 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.041708 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.041735 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.041747 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.041762 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.041771 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:24Z","lastTransitionTime":"2025-11-23T06:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.062520 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:24Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.143516 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.143549 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.143558 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.143572 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.143582 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:24Z","lastTransitionTime":"2025-11-23T06:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.245667 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.245714 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.245723 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.245736 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.245744 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:24Z","lastTransitionTime":"2025-11-23T06:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.273116 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.273183 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:24 crc kubenswrapper[4559]: E1123 06:45:24.273256 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:45:24 crc kubenswrapper[4559]: E1123 06:45:24.273398 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.347403 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.347431 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.347439 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.347452 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.347461 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:24Z","lastTransitionTime":"2025-11-23T06:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.371548 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-w5tlv" event={"ID":"cdf91a44-8c01-4104-8d6f-e76b8c321166","Type":"ContainerStarted","Data":"e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78"} Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.373224 4559 generic.go:334] "Generic (PLEG): container finished" podID="e9164873-0a60-498c-9b37-cd73817b066a" containerID="8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74" exitCode=0 Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.373259 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" event={"ID":"e9164873-0a60-498c-9b37-cd73817b066a","Type":"ContainerDied","Data":"8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74"} Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.381503 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:24Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.391443 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:24Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.400021 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:24Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.409153 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:24Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.418554 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:24Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.427605 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:24Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.435455 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:24Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.445835 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:24Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.448918 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.448953 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.448963 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.448976 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.448985 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:24Z","lastTransitionTime":"2025-11-23T06:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.454156 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:24Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.463885 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:24Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.503482 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:24Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.547852 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:24Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.551013 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.551040 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.551049 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.551063 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.551073 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:24Z","lastTransitionTime":"2025-11-23T06:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.583693 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:24Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.623470 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:24Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.652497 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.652529 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.652539 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.652551 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.652560 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:24Z","lastTransitionTime":"2025-11-23T06:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.663569 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:24Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.704800 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:24Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.745750 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:24Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.754135 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.754164 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.754174 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.754188 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.754196 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:24Z","lastTransitionTime":"2025-11-23T06:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.785056 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:24Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.823983 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:24Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.856467 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.856504 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.856514 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.856533 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.856545 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:24Z","lastTransitionTime":"2025-11-23T06:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.865007 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:24Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.904112 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:24Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.944378 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:24Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.959031 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.959058 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.959066 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.959080 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.959089 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:24Z","lastTransitionTime":"2025-11-23T06:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:24 crc kubenswrapper[4559]: I1123 06:45:24.987293 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:24Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.023974 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.061421 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.061452 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.061461 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.061503 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.061511 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:25Z","lastTransitionTime":"2025-11-23T06:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.063410 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.105192 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.143836 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.163625 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.163681 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.163690 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.163705 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.163714 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:25Z","lastTransitionTime":"2025-11-23T06:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.183365 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.265423 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.265450 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.265459 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.265474 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.265484 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:25Z","lastTransitionTime":"2025-11-23T06:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.273149 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:25 crc kubenswrapper[4559]: E1123 06:45:25.273268 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.367320 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.367358 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.367367 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.367383 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.367392 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:25Z","lastTransitionTime":"2025-11-23T06:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.376885 4559 generic.go:334] "Generic (PLEG): container finished" podID="e9164873-0a60-498c-9b37-cd73817b066a" containerID="7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be" exitCode=0 Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.376946 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" event={"ID":"e9164873-0a60-498c-9b37-cd73817b066a","Type":"ContainerDied","Data":"7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be"} Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.379889 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" event={"ID":"b12e257e-7887-4795-9221-4db8fd5856dd","Type":"ContainerStarted","Data":"52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90"} Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.385603 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.393784 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.408026 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.416251 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.422696 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.431267 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.463439 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.469807 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.469831 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.469840 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.469855 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.469863 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:25Z","lastTransitionTime":"2025-11-23T06:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.502529 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.545289 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.571626 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.571681 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.571689 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.571704 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.571712 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:25Z","lastTransitionTime":"2025-11-23T06:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.585072 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.657746 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.670125 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.673565 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.673596 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.673605 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.673620 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.673628 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:25Z","lastTransitionTime":"2025-11-23T06:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.707444 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.746604 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:25Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.775352 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.775381 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.775390 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.775400 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.775408 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:25Z","lastTransitionTime":"2025-11-23T06:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.877405 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.877434 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.877443 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.877454 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.877462 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:25Z","lastTransitionTime":"2025-11-23T06:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.978763 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.978791 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.978800 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.978812 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:25 crc kubenswrapper[4559]: I1123 06:45:25.978821 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:25Z","lastTransitionTime":"2025-11-23T06:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.080904 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.080931 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.080940 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.080953 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.080961 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:26Z","lastTransitionTime":"2025-11-23T06:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.182635 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.182705 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.182714 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.182725 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.182733 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:26Z","lastTransitionTime":"2025-11-23T06:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.273072 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.273127 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:26 crc kubenswrapper[4559]: E1123 06:45:26.273195 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:45:26 crc kubenswrapper[4559]: E1123 06:45:26.273300 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.284338 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.284363 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.284372 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.284383 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.284392 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:26Z","lastTransitionTime":"2025-11-23T06:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.384817 4559 generic.go:334] "Generic (PLEG): container finished" podID="e9164873-0a60-498c-9b37-cd73817b066a" containerID="878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36" exitCode=0 Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.384854 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" event={"ID":"e9164873-0a60-498c-9b37-cd73817b066a","Type":"ContainerDied","Data":"878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36"} Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.385790 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.385805 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.385813 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.385823 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.385830 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:26Z","lastTransitionTime":"2025-11-23T06:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.393368 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:26Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.404759 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:26Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.415691 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:26Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.425214 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:26Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.434530 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:26Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.442789 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:26Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.450359 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:26Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.458245 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:26Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.470086 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:26Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.480904 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:26Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.487203 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.487225 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.487233 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.487246 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.487254 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:26Z","lastTransitionTime":"2025-11-23T06:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.488147 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:26Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.497266 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:26Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.505101 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:26Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.512564 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:26Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.589108 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.589136 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.589145 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.589155 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.589164 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:26Z","lastTransitionTime":"2025-11-23T06:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.691276 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.691306 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.691316 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.691327 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.691335 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:26Z","lastTransitionTime":"2025-11-23T06:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.793228 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.793256 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.793265 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.793274 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.793282 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:26Z","lastTransitionTime":"2025-11-23T06:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.895185 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.895807 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.895823 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.895836 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.895845 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:26Z","lastTransitionTime":"2025-11-23T06:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.997446 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.997478 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.997486 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.997500 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:26 crc kubenswrapper[4559]: I1123 06:45:26.997514 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:26Z","lastTransitionTime":"2025-11-23T06:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.099305 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.099335 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.099344 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.099357 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.099387 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:27Z","lastTransitionTime":"2025-11-23T06:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.201211 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.201244 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.201252 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.201265 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.201273 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:27Z","lastTransitionTime":"2025-11-23T06:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.273430 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:27 crc kubenswrapper[4559]: E1123 06:45:27.273521 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.303339 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.303366 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.303373 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.303384 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.303394 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:27Z","lastTransitionTime":"2025-11-23T06:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.389751 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" event={"ID":"e9164873-0a60-498c-9b37-cd73817b066a","Type":"ContainerStarted","Data":"56e33b4cd5df561efe04dcbbf6688c9edca01209b83ca44b40fc61cf562b2e82"} Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.392371 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" event={"ID":"b12e257e-7887-4795-9221-4db8fd5856dd","Type":"ContainerStarted","Data":"d23600fbbee9a1d284c8e196e82b34efc7d9be09068461d89f2b6a43225c7f13"} Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.392524 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.392547 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.400923 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:27Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.404970 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.404995 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.405004 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.405015 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.405024 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:27Z","lastTransitionTime":"2025-11-23T06:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.410158 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.410199 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.412520 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:27Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.419873 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:27Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.428524 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:27Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.435975 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:27Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.443326 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:27Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.450623 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:27Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.458930 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:27Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.467945 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56e33b4cd5df561efe04dcbbf6688c9edca01209b83ca44b40fc61cf562b2e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:27Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.476349 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:27Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.483119 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:27Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.494860 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:27Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.502691 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:27Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.507255 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.507284 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.507291 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.507303 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.507311 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:27Z","lastTransitionTime":"2025-11-23T06:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.509117 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:27Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.518899 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:27Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.525618 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:27Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.537428 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d23600fbbee9a1d284c8e196e82b34efc7d9be09068461d89f2b6a43225c7f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:27Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.545576 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:27Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.551699 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:27Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.560394 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:27Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.568931 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:27Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.575210 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:27Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.583040 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:27Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.590737 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:27Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.599499 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56e33b4cd5df561efe04dcbbf6688c9edca01209b83ca44b40fc61cf562b2e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:27Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.608057 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:27Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.608859 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.608883 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.608891 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.608903 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.608911 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:27Z","lastTransitionTime":"2025-11-23T06:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.615671 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:27Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.623395 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:27Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.710912 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.710936 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.710946 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.710956 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.710962 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:27Z","lastTransitionTime":"2025-11-23T06:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.812376 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.812396 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.812404 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.812413 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.812421 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:27Z","lastTransitionTime":"2025-11-23T06:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.914060 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.914090 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.914098 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.914112 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:27 crc kubenswrapper[4559]: I1123 06:45:27.914121 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:27Z","lastTransitionTime":"2025-11-23T06:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.015944 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.015966 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.015975 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.015986 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.015994 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:28Z","lastTransitionTime":"2025-11-23T06:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.116873 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.116899 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.116908 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.116918 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.116926 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:28Z","lastTransitionTime":"2025-11-23T06:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.219290 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.219324 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.219332 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.219346 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.219355 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:28Z","lastTransitionTime":"2025-11-23T06:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.273309 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.273356 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:28 crc kubenswrapper[4559]: E1123 06:45:28.273413 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:45:28 crc kubenswrapper[4559]: E1123 06:45:28.273678 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.283738 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.290351 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.299882 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.306421 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.316682 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.321024 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.321061 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.321072 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.321087 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.321095 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:28Z","lastTransitionTime":"2025-11-23T06:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.326436 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.336854 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.355565 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.370683 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.382875 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56e33b4cd5df561efe04dcbbf6688c9edca01209b83ca44b40fc61cf562b2e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.392495 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.394270 4559 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.400065 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.412189 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d23600fbbee9a1d284c8e196e82b34efc7d9be09068461d89f2b6a43225c7f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.420126 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.423367 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.423395 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.423403 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.423415 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.423424 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:28Z","lastTransitionTime":"2025-11-23T06:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.525135 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.525167 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.525175 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.525188 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.525195 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:28Z","lastTransitionTime":"2025-11-23T06:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.626854 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.626883 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.626892 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.626903 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.626912 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:28Z","lastTransitionTime":"2025-11-23T06:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.728698 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.728903 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.728912 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.728923 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.728932 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:28Z","lastTransitionTime":"2025-11-23T06:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.787567 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.787597 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.787632 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.787656 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.787666 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:28Z","lastTransitionTime":"2025-11-23T06:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:28 crc kubenswrapper[4559]: E1123 06:45:28.797002 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.799626 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.799713 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.799750 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.799766 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.799780 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:28Z","lastTransitionTime":"2025-11-23T06:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:28 crc kubenswrapper[4559]: E1123 06:45:28.808591 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.810916 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.810945 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.810955 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.810970 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.810978 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:28Z","lastTransitionTime":"2025-11-23T06:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:28 crc kubenswrapper[4559]: E1123 06:45:28.819113 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.821472 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.821504 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.821513 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.821526 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.821537 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:28Z","lastTransitionTime":"2025-11-23T06:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:28 crc kubenswrapper[4559]: E1123 06:45:28.829046 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.831359 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.831382 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.831390 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.831400 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.831408 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:28Z","lastTransitionTime":"2025-11-23T06:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:28 crc kubenswrapper[4559]: E1123 06:45:28.839067 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:28 crc kubenswrapper[4559]: E1123 06:45:28.839171 4559 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.840255 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.840283 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.840292 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.840302 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.840309 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:28Z","lastTransitionTime":"2025-11-23T06:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.941987 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.942015 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.942023 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.942036 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:28 crc kubenswrapper[4559]: I1123 06:45:28.942044 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:28Z","lastTransitionTime":"2025-11-23T06:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.043972 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.043997 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.044005 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.044015 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.044022 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:29Z","lastTransitionTime":"2025-11-23T06:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.145324 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.145347 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.145354 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.145366 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.145375 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:29Z","lastTransitionTime":"2025-11-23T06:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.246563 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.246603 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.246627 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.246660 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.246672 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:29Z","lastTransitionTime":"2025-11-23T06:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.272944 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:29 crc kubenswrapper[4559]: E1123 06:45:29.273027 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.348709 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.348739 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.348747 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.348758 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.348770 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:29Z","lastTransitionTime":"2025-11-23T06:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.397361 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jt56f_b12e257e-7887-4795-9221-4db8fd5856dd/ovnkube-controller/0.log" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.399216 4559 generic.go:334] "Generic (PLEG): container finished" podID="b12e257e-7887-4795-9221-4db8fd5856dd" containerID="d23600fbbee9a1d284c8e196e82b34efc7d9be09068461d89f2b6a43225c7f13" exitCode=1 Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.399249 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" event={"ID":"b12e257e-7887-4795-9221-4db8fd5856dd","Type":"ContainerDied","Data":"d23600fbbee9a1d284c8e196e82b34efc7d9be09068461d89f2b6a43225c7f13"} Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.399764 4559 scope.go:117] "RemoveContainer" containerID="d23600fbbee9a1d284c8e196e82b34efc7d9be09068461d89f2b6a43225c7f13" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.408188 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.416536 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.428384 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d23600fbbee9a1d284c8e196e82b34efc7d9be09068461d89f2b6a43225c7f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d23600fbbee9a1d284c8e196e82b34efc7d9be09068461d89f2b6a43225c7f13\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1123 06:45:28.789273 5898 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1123 06:45:28.789323 5898 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1123 06:45:28.789350 5898 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1123 06:45:28.789387 5898 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1123 06:45:28.789417 5898 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1123 06:45:28.789439 5898 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1123 06:45:28.789458 5898 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1123 06:45:28.789525 5898 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1123 06:45:28.789582 5898 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1123 06:45:28.789470 5898 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1123 06:45:28.789570 5898 handler.go:208] Removed *v1.Node event handler 7\\\\nI1123 06:45:28.789632 5898 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1123 06:45:28.789649 5898 handler.go:208] Removed *v1.Node event handler 2\\\\nI1123 06:45:28.789656 5898 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1123 06:45:28.789684 5898 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.436455 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.443109 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.450269 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.450295 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.450304 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.450315 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.450324 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:29Z","lastTransitionTime":"2025-11-23T06:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.451690 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.459455 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.465709 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.473269 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.481100 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.494135 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56e33b4cd5df561efe04dcbbf6688c9edca01209b83ca44b40fc61cf562b2e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.507919 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.515871 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.524208 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.552189 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.552217 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.552226 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.552240 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.552249 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:29Z","lastTransitionTime":"2025-11-23T06:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.653530 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.653564 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.653573 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.653590 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.653599 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:29Z","lastTransitionTime":"2025-11-23T06:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.745192 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.754052 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.754962 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.754993 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.755003 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.755014 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.755022 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:29Z","lastTransitionTime":"2025-11-23T06:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.761635 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.768536 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.783152 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.794463 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.804888 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.812662 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.821562 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.830953 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56e33b4cd5df561efe04dcbbf6688c9edca01209b83ca44b40fc61cf562b2e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.840367 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.848757 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.857140 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.857170 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.857179 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.857191 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.857200 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:29Z","lastTransitionTime":"2025-11-23T06:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.861481 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d23600fbbee9a1d284c8e196e82b34efc7d9be09068461d89f2b6a43225c7f13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d23600fbbee9a1d284c8e196e82b34efc7d9be09068461d89f2b6a43225c7f13\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1123 06:45:28.789273 5898 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1123 06:45:28.789323 5898 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1123 06:45:28.789350 5898 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1123 06:45:28.789387 5898 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1123 06:45:28.789417 5898 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1123 06:45:28.789439 5898 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1123 06:45:28.789458 5898 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1123 06:45:28.789525 5898 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1123 06:45:28.789582 5898 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1123 06:45:28.789470 5898 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1123 06:45:28.789570 5898 handler.go:208] Removed *v1.Node event handler 7\\\\nI1123 06:45:28.789632 5898 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1123 06:45:28.789649 5898 handler.go:208] Removed *v1.Node event handler 2\\\\nI1123 06:45:28.789656 5898 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1123 06:45:28.789684 5898 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.869145 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.876082 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:29Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.958559 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.958586 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.958594 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.958615 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:29 crc kubenswrapper[4559]: I1123 06:45:29.958625 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:29Z","lastTransitionTime":"2025-11-23T06:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.060307 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.060341 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.060351 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.060364 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.060372 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:30Z","lastTransitionTime":"2025-11-23T06:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.164328 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.164378 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.164388 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.164403 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.164412 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:30Z","lastTransitionTime":"2025-11-23T06:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.266186 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.266210 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.266219 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.266229 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.266238 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:30Z","lastTransitionTime":"2025-11-23T06:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.273391 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.273414 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:30 crc kubenswrapper[4559]: E1123 06:45:30.273489 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:45:30 crc kubenswrapper[4559]: E1123 06:45:30.273624 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.368217 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.368246 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.368255 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.368267 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.368276 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:30Z","lastTransitionTime":"2025-11-23T06:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.402377 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jt56f_b12e257e-7887-4795-9221-4db8fd5856dd/ovnkube-controller/1.log" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.402808 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jt56f_b12e257e-7887-4795-9221-4db8fd5856dd/ovnkube-controller/0.log" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.404706 4559 generic.go:334] "Generic (PLEG): container finished" podID="b12e257e-7887-4795-9221-4db8fd5856dd" containerID="d455758b0f3b64148d612aca66c657881d20868d079f7d59ddc147936982e6d3" exitCode=1 Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.404732 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" event={"ID":"b12e257e-7887-4795-9221-4db8fd5856dd","Type":"ContainerDied","Data":"d455758b0f3b64148d612aca66c657881d20868d079f7d59ddc147936982e6d3"} Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.404763 4559 scope.go:117] "RemoveContainer" containerID="d23600fbbee9a1d284c8e196e82b34efc7d9be09068461d89f2b6a43225c7f13" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.405179 4559 scope.go:117] "RemoveContainer" containerID="d455758b0f3b64148d612aca66c657881d20868d079f7d59ddc147936982e6d3" Nov 23 06:45:30 crc kubenswrapper[4559]: E1123 06:45:30.405348 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jt56f_openshift-ovn-kubernetes(b12e257e-7887-4795-9221-4db8fd5856dd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.415979 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.423410 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.431551 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.439376 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.447885 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.455042 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.462862 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.470120 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.470148 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.470157 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.470168 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.470176 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:30Z","lastTransitionTime":"2025-11-23T06:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.472452 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56e33b4cd5df561efe04dcbbf6688c9edca01209b83ca44b40fc61cf562b2e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.480815 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.487487 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.498920 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d455758b0f3b64148d612aca66c657881d20868d079f7d59ddc147936982e6d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d23600fbbee9a1d284c8e196e82b34efc7d9be09068461d89f2b6a43225c7f13\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:45:28Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1123 06:45:28.789273 5898 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1123 06:45:28.789323 5898 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1123 06:45:28.789350 5898 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1123 06:45:28.789387 5898 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1123 06:45:28.789417 5898 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1123 06:45:28.789439 5898 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1123 06:45:28.789458 5898 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1123 06:45:28.789525 5898 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1123 06:45:28.789582 5898 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1123 06:45:28.789470 5898 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1123 06:45:28.789570 5898 handler.go:208] Removed *v1.Node event handler 7\\\\nI1123 06:45:28.789632 5898 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1123 06:45:28.789649 5898 handler.go:208] Removed *v1.Node event handler 2\\\\nI1123 06:45:28.789656 5898 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1123 06:45:28.789684 5898 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d455758b0f3b64148d612aca66c657881d20868d079f7d59ddc147936982e6d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"message\\\":\\\"-machine-approver/machine-approver. OVN-Kubernetes controller took 0.103676017 seconds. No OVN measurement.\\\\nI1123 06:45:29.983899 6024 services_controller.go:445] Built service openshift-kube-apiserver/apiserver LB template configs for network=default: []services.lbConfig(nil)\\\\nI1123 06:45:29.983901 6024 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-s7zkt in node crc\\\\nI1123 06:45:29.983906 6024 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-s7zkt after 0 failed attempt(s)\\\\nI1123 06:45:29.983905 6024 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-rlxnw\\\\nI1123 06:45:29.983910 6024 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-apiserver/apiserver\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.93\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.506475 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.513793 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.520070 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.571472 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.571499 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.571507 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.571519 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.571528 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:30Z","lastTransitionTime":"2025-11-23T06:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.673275 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.673302 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.673309 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.673321 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.673330 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:30Z","lastTransitionTime":"2025-11-23T06:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.775092 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.775116 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.775125 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.775134 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.775142 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:30Z","lastTransitionTime":"2025-11-23T06:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.876374 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.876403 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.876412 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.876422 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.876429 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:30Z","lastTransitionTime":"2025-11-23T06:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.977534 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.977580 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.977611 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.977631 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:30 crc kubenswrapper[4559]: I1123 06:45:30.977667 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:30Z","lastTransitionTime":"2025-11-23T06:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.079200 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.079230 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.079240 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.079252 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.079262 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:31Z","lastTransitionTime":"2025-11-23T06:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.180516 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.180538 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.180546 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.180560 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.180569 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:31Z","lastTransitionTime":"2025-11-23T06:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.273231 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:31 crc kubenswrapper[4559]: E1123 06:45:31.273315 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.281693 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.281796 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.281804 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.281815 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.281823 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:31Z","lastTransitionTime":"2025-11-23T06:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.383434 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.383462 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.383471 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.383482 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.383490 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:31Z","lastTransitionTime":"2025-11-23T06:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.410051 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jt56f_b12e257e-7887-4795-9221-4db8fd5856dd/ovnkube-controller/1.log" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.412662 4559 scope.go:117] "RemoveContainer" containerID="d455758b0f3b64148d612aca66c657881d20868d079f7d59ddc147936982e6d3" Nov 23 06:45:31 crc kubenswrapper[4559]: E1123 06:45:31.412777 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jt56f_openshift-ovn-kubernetes(b12e257e-7887-4795-9221-4db8fd5856dd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.421038 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.427316 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.435073 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.441855 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.450606 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.458500 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.466348 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.473239 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.480849 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.485518 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.485552 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.485561 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.485571 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.485579 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:31Z","lastTransitionTime":"2025-11-23T06:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.491609 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56e33b4cd5df561efe04dcbbf6688c9edca01209b83ca44b40fc61cf562b2e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.500302 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.507049 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.518285 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d455758b0f3b64148d612aca66c657881d20868d079f7d59ddc147936982e6d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d455758b0f3b64148d612aca66c657881d20868d079f7d59ddc147936982e6d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"message\\\":\\\"-machine-approver/machine-approver. OVN-Kubernetes controller took 0.103676017 seconds. No OVN measurement.\\\\nI1123 06:45:29.983899 6024 services_controller.go:445] Built service openshift-kube-apiserver/apiserver LB template configs for network=default: []services.lbConfig(nil)\\\\nI1123 06:45:29.983901 6024 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-s7zkt in node crc\\\\nI1123 06:45:29.983906 6024 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-s7zkt after 0 failed attempt(s)\\\\nI1123 06:45:29.983905 6024 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-rlxnw\\\\nI1123 06:45:29.983910 6024 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-apiserver/apiserver\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.93\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jt56f_openshift-ovn-kubernetes(b12e257e-7887-4795-9221-4db8fd5856dd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.525848 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:31Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.587213 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.587247 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.587256 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.587269 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.587277 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:31Z","lastTransitionTime":"2025-11-23T06:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.689344 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.689370 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.689378 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.689388 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.689397 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:31Z","lastTransitionTime":"2025-11-23T06:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.791238 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.791288 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.791296 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.791308 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.791317 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:31Z","lastTransitionTime":"2025-11-23T06:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.893008 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.893046 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.893054 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.893067 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.893077 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:31Z","lastTransitionTime":"2025-11-23T06:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.900318 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:31 crc kubenswrapper[4559]: E1123 06:45:31.900441 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:45:31 crc kubenswrapper[4559]: E1123 06:45:31.900463 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:45:31 crc kubenswrapper[4559]: E1123 06:45:31.900473 4559 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:45:31 crc kubenswrapper[4559]: E1123 06:45:31.900512 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-23 06:45:47.900500798 +0000 UTC m=+49.922486401 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.994785 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.994816 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.994824 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.994836 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:31 crc kubenswrapper[4559]: I1123 06:45:31.994846 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:31Z","lastTransitionTime":"2025-11-23T06:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.001029 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.001143 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:32 crc kubenswrapper[4559]: E1123 06:45:32.001153 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:45:48.001140854 +0000 UTC m=+50.023126469 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:45:32 crc kubenswrapper[4559]: E1123 06:45:32.001235 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:45:32 crc kubenswrapper[4559]: E1123 06:45:32.001250 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:45:32 crc kubenswrapper[4559]: E1123 06:45:32.001259 4559 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:45:32 crc kubenswrapper[4559]: E1123 06:45:32.001298 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-23 06:45:48.001287176 +0000 UTC m=+50.023272791 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.096853 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.096884 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.096893 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.096904 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.096913 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:32Z","lastTransitionTime":"2025-11-23T06:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.102305 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.102342 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:32 crc kubenswrapper[4559]: E1123 06:45:32.102428 4559 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:45:32 crc kubenswrapper[4559]: E1123 06:45:32.102470 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:45:48.102461842 +0000 UTC m=+50.124447455 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:45:32 crc kubenswrapper[4559]: E1123 06:45:32.102480 4559 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:45:32 crc kubenswrapper[4559]: E1123 06:45:32.102537 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:45:48.102523416 +0000 UTC m=+50.124509030 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.198310 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.198333 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.198342 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.198387 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.198399 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:32Z","lastTransitionTime":"2025-11-23T06:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.273319 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:32 crc kubenswrapper[4559]: E1123 06:45:32.273425 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.273328 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:32 crc kubenswrapper[4559]: E1123 06:45:32.273501 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.299927 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.299949 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.299958 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.299968 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.299976 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:32Z","lastTransitionTime":"2025-11-23T06:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.318325 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh"] Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.318628 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.320304 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.320857 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.327410 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:32Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.334181 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:32Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.341131 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5bf62d2-3296-4245-970b-2553e647234c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9k8xh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:32Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.349859 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:32Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.357338 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:32Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.363600 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:32Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.373613 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:32Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.381740 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:32Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.391466 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:32Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.399236 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:32Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.401446 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.401476 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.401486 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.401504 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.401513 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:32Z","lastTransitionTime":"2025-11-23T06:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.404168 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a5bf62d2-3296-4245-970b-2553e647234c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-9k8xh\" (UID: \"a5bf62d2-3296-4245-970b-2553e647234c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.404220 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl8jl\" (UniqueName: \"kubernetes.io/projected/a5bf62d2-3296-4245-970b-2553e647234c-kube-api-access-wl8jl\") pod \"ovnkube-control-plane-749d76644c-9k8xh\" (UID: \"a5bf62d2-3296-4245-970b-2553e647234c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.404243 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a5bf62d2-3296-4245-970b-2553e647234c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-9k8xh\" (UID: \"a5bf62d2-3296-4245-970b-2553e647234c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.404297 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a5bf62d2-3296-4245-970b-2553e647234c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-9k8xh\" (UID: \"a5bf62d2-3296-4245-970b-2553e647234c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.407839 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:32Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.416857 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56e33b4cd5df561efe04dcbbf6688c9edca01209b83ca44b40fc61cf562b2e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:32Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.424624 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:32Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.431399 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:32Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.443458 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d455758b0f3b64148d612aca66c657881d20868d079f7d59ddc147936982e6d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d455758b0f3b64148d612aca66c657881d20868d079f7d59ddc147936982e6d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"message\\\":\\\"-machine-approver/machine-approver. OVN-Kubernetes controller took 0.103676017 seconds. No OVN measurement.\\\\nI1123 06:45:29.983899 6024 services_controller.go:445] Built service openshift-kube-apiserver/apiserver LB template configs for network=default: []services.lbConfig(nil)\\\\nI1123 06:45:29.983901 6024 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-s7zkt in node crc\\\\nI1123 06:45:29.983906 6024 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-s7zkt after 0 failed attempt(s)\\\\nI1123 06:45:29.983905 6024 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-rlxnw\\\\nI1123 06:45:29.983910 6024 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-apiserver/apiserver\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.93\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jt56f_openshift-ovn-kubernetes(b12e257e-7887-4795-9221-4db8fd5856dd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:32Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.503755 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.503788 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.503798 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.503815 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.503825 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:32Z","lastTransitionTime":"2025-11-23T06:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.505037 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl8jl\" (UniqueName: \"kubernetes.io/projected/a5bf62d2-3296-4245-970b-2553e647234c-kube-api-access-wl8jl\") pod \"ovnkube-control-plane-749d76644c-9k8xh\" (UID: \"a5bf62d2-3296-4245-970b-2553e647234c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.505078 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a5bf62d2-3296-4245-970b-2553e647234c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-9k8xh\" (UID: \"a5bf62d2-3296-4245-970b-2553e647234c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.505103 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a5bf62d2-3296-4245-970b-2553e647234c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-9k8xh\" (UID: \"a5bf62d2-3296-4245-970b-2553e647234c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.505121 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a5bf62d2-3296-4245-970b-2553e647234c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-9k8xh\" (UID: \"a5bf62d2-3296-4245-970b-2553e647234c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.505627 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a5bf62d2-3296-4245-970b-2553e647234c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-9k8xh\" (UID: \"a5bf62d2-3296-4245-970b-2553e647234c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.506050 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a5bf62d2-3296-4245-970b-2553e647234c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-9k8xh\" (UID: \"a5bf62d2-3296-4245-970b-2553e647234c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.508757 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a5bf62d2-3296-4245-970b-2553e647234c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-9k8xh\" (UID: \"a5bf62d2-3296-4245-970b-2553e647234c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.516905 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl8jl\" (UniqueName: \"kubernetes.io/projected/a5bf62d2-3296-4245-970b-2553e647234c-kube-api-access-wl8jl\") pod \"ovnkube-control-plane-749d76644c-9k8xh\" (UID: \"a5bf62d2-3296-4245-970b-2553e647234c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.605779 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.605810 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.605820 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.605831 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.605839 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:32Z","lastTransitionTime":"2025-11-23T06:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.626958 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.707977 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.708003 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.708011 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.708023 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.708032 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:32Z","lastTransitionTime":"2025-11-23T06:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.809312 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.809342 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.809352 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.809365 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.809373 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:32Z","lastTransitionTime":"2025-11-23T06:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.910845 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.910877 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.910885 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.910896 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:32 crc kubenswrapper[4559]: I1123 06:45:32.910905 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:32Z","lastTransitionTime":"2025-11-23T06:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.012481 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.012709 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.012799 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.012863 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.012913 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:33Z","lastTransitionTime":"2025-11-23T06:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.115250 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.115280 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.115288 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.115299 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.115306 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:33Z","lastTransitionTime":"2025-11-23T06:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.217550 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.217786 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.217855 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.217956 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.218021 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:33Z","lastTransitionTime":"2025-11-23T06:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.273192 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:33 crc kubenswrapper[4559]: E1123 06:45:33.273361 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.319365 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.319405 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.319414 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.319429 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.319440 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:33Z","lastTransitionTime":"2025-11-23T06:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.418129 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" event={"ID":"a5bf62d2-3296-4245-970b-2553e647234c","Type":"ContainerStarted","Data":"cbebf00280a1ffac1595cd96655356e8567fe98e96ab346e0225f2fcc1c4de9c"} Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.418174 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" event={"ID":"a5bf62d2-3296-4245-970b-2553e647234c","Type":"ContainerStarted","Data":"a5eb007f4dc7b9bbbed98a8780b4a9ceed7fca2ccd970acd5f7f8a734ae61c55"} Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.418186 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" event={"ID":"a5bf62d2-3296-4245-970b-2553e647234c","Type":"ContainerStarted","Data":"1d6cb58d6c9cd88e36de9faf45996bc7e09ba7f5dbdfafb603c4d5257f1e5da5"} Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.420882 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.420912 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.420921 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.420934 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.420943 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:33Z","lastTransitionTime":"2025-11-23T06:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.428548 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.437018 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.444558 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.452697 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.461393 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56e33b4cd5df561efe04dcbbf6688c9edca01209b83ca44b40fc61cf562b2e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.470742 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.478333 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.486332 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.494239 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.501958 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.511856 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.522842 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.522876 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.522885 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.522898 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.522907 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:33Z","lastTransitionTime":"2025-11-23T06:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.525191 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d455758b0f3b64148d612aca66c657881d20868d079f7d59ddc147936982e6d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d455758b0f3b64148d612aca66c657881d20868d079f7d59ddc147936982e6d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"message\\\":\\\"-machine-approver/machine-approver. OVN-Kubernetes controller took 0.103676017 seconds. No OVN measurement.\\\\nI1123 06:45:29.983899 6024 services_controller.go:445] Built service openshift-kube-apiserver/apiserver LB template configs for network=default: []services.lbConfig(nil)\\\\nI1123 06:45:29.983901 6024 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-s7zkt in node crc\\\\nI1123 06:45:29.983906 6024 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-s7zkt after 0 failed attempt(s)\\\\nI1123 06:45:29.983905 6024 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-rlxnw\\\\nI1123 06:45:29.983910 6024 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-apiserver/apiserver\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.93\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jt56f_openshift-ovn-kubernetes(b12e257e-7887-4795-9221-4db8fd5856dd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.533750 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.539925 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.547453 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5bf62d2-3296-4245-970b-2553e647234c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5eb007f4dc7b9bbbed98a8780b4a9ceed7fca2ccd970acd5f7f8a734ae61c55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbebf00280a1ffac1595cd96655356e8567fe98e96ab346e0225f2fcc1c4de9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9k8xh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.625127 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.625158 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.625167 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.625178 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.625187 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:33Z","lastTransitionTime":"2025-11-23T06:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.727301 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.727335 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.727343 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.727356 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.727364 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:33Z","lastTransitionTime":"2025-11-23T06:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.733298 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-svb6z"] Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.733682 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:45:33 crc kubenswrapper[4559]: E1123 06:45:33.733727 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.741768 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.748689 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.760138 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d455758b0f3b64148d612aca66c657881d20868d079f7d59ddc147936982e6d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d455758b0f3b64148d612aca66c657881d20868d079f7d59ddc147936982e6d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"message\\\":\\\"-machine-approver/machine-approver. OVN-Kubernetes controller took 0.103676017 seconds. No OVN measurement.\\\\nI1123 06:45:29.983899 6024 services_controller.go:445] Built service openshift-kube-apiserver/apiserver LB template configs for network=default: []services.lbConfig(nil)\\\\nI1123 06:45:29.983901 6024 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-s7zkt in node crc\\\\nI1123 06:45:29.983906 6024 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-s7zkt after 0 failed attempt(s)\\\\nI1123 06:45:29.983905 6024 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-rlxnw\\\\nI1123 06:45:29.983910 6024 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-apiserver/apiserver\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.93\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jt56f_openshift-ovn-kubernetes(b12e257e-7887-4795-9221-4db8fd5856dd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.768012 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.774238 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.781307 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5bf62d2-3296-4245-970b-2553e647234c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5eb007f4dc7b9bbbed98a8780b4a9ceed7fca2ccd970acd5f7f8a734ae61c55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbebf00280a1ffac1595cd96655356e8567fe98e96ab346e0225f2fcc1c4de9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9k8xh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.789321 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.796713 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.803016 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.812288 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56e33b4cd5df561efe04dcbbf6688c9edca01209b83ca44b40fc61cf562b2e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.816194 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4xp4\" (UniqueName: \"kubernetes.io/projected/abb6827e-709d-4ea9-8090-e4f541b1ff3f-kube-api-access-r4xp4\") pod \"network-metrics-daemon-svb6z\" (UID: \"abb6827e-709d-4ea9-8090-e4f541b1ff3f\") " pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.816227 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/abb6827e-709d-4ea9-8090-e4f541b1ff3f-metrics-certs\") pod \"network-metrics-daemon-svb6z\" (UID: \"abb6827e-709d-4ea9-8090-e4f541b1ff3f\") " pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.819790 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svb6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abb6827e-709d-4ea9-8090-e4f541b1ff3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svb6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.828569 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.829094 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.829123 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.829131 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.829143 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.829153 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:33Z","lastTransitionTime":"2025-11-23T06:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.836312 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.843710 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.850448 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.858282 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:33Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.916891 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4xp4\" (UniqueName: \"kubernetes.io/projected/abb6827e-709d-4ea9-8090-e4f541b1ff3f-kube-api-access-r4xp4\") pod \"network-metrics-daemon-svb6z\" (UID: \"abb6827e-709d-4ea9-8090-e4f541b1ff3f\") " pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.916929 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/abb6827e-709d-4ea9-8090-e4f541b1ff3f-metrics-certs\") pod \"network-metrics-daemon-svb6z\" (UID: \"abb6827e-709d-4ea9-8090-e4f541b1ff3f\") " pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:45:33 crc kubenswrapper[4559]: E1123 06:45:33.917039 4559 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:45:33 crc kubenswrapper[4559]: E1123 06:45:33.917076 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/abb6827e-709d-4ea9-8090-e4f541b1ff3f-metrics-certs podName:abb6827e-709d-4ea9-8090-e4f541b1ff3f nodeName:}" failed. No retries permitted until 2025-11-23 06:45:34.417065795 +0000 UTC m=+36.439051409 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/abb6827e-709d-4ea9-8090-e4f541b1ff3f-metrics-certs") pod "network-metrics-daemon-svb6z" (UID: "abb6827e-709d-4ea9-8090-e4f541b1ff3f") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.930090 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4xp4\" (UniqueName: \"kubernetes.io/projected/abb6827e-709d-4ea9-8090-e4f541b1ff3f-kube-api-access-r4xp4\") pod \"network-metrics-daemon-svb6z\" (UID: \"abb6827e-709d-4ea9-8090-e4f541b1ff3f\") " pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.931030 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.931049 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.931058 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.931067 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:33 crc kubenswrapper[4559]: I1123 06:45:33.931074 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:33Z","lastTransitionTime":"2025-11-23T06:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.033015 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.033041 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.033050 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.033061 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.033068 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:34Z","lastTransitionTime":"2025-11-23T06:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.135145 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.135393 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.135471 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.135532 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.135607 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:34Z","lastTransitionTime":"2025-11-23T06:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.237230 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.237256 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.237264 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.237275 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.237283 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:34Z","lastTransitionTime":"2025-11-23T06:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.273092 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.273113 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:34 crc kubenswrapper[4559]: E1123 06:45:34.273177 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:45:34 crc kubenswrapper[4559]: E1123 06:45:34.273237 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.338297 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.338472 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.338544 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.338626 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.338729 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:34Z","lastTransitionTime":"2025-11-23T06:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.421003 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/abb6827e-709d-4ea9-8090-e4f541b1ff3f-metrics-certs\") pod \"network-metrics-daemon-svb6z\" (UID: \"abb6827e-709d-4ea9-8090-e4f541b1ff3f\") " pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:45:34 crc kubenswrapper[4559]: E1123 06:45:34.421451 4559 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:45:34 crc kubenswrapper[4559]: E1123 06:45:34.421571 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/abb6827e-709d-4ea9-8090-e4f541b1ff3f-metrics-certs podName:abb6827e-709d-4ea9-8090-e4f541b1ff3f nodeName:}" failed. No retries permitted until 2025-11-23 06:45:35.421546546 +0000 UTC m=+37.443532160 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/abb6827e-709d-4ea9-8090-e4f541b1ff3f-metrics-certs") pod "network-metrics-daemon-svb6z" (UID: "abb6827e-709d-4ea9-8090-e4f541b1ff3f") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.440693 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.440720 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.440729 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.440739 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.440746 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:34Z","lastTransitionTime":"2025-11-23T06:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.542286 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.542314 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.542322 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.542332 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.542340 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:34Z","lastTransitionTime":"2025-11-23T06:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.644629 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.644690 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.644700 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.644717 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.644730 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:34Z","lastTransitionTime":"2025-11-23T06:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.746306 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.746345 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.746353 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.746368 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.746376 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:34Z","lastTransitionTime":"2025-11-23T06:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.847871 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.847902 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.847911 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.847924 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.847933 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:34Z","lastTransitionTime":"2025-11-23T06:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.950238 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.950266 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.950274 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.950285 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:34 crc kubenswrapper[4559]: I1123 06:45:34.950294 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:34Z","lastTransitionTime":"2025-11-23T06:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.051571 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.051602 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.051613 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.051626 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.051634 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:35Z","lastTransitionTime":"2025-11-23T06:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.155002 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.155038 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.155046 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.155057 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.155066 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:35Z","lastTransitionTime":"2025-11-23T06:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.256608 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.256633 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.256662 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.256674 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.256681 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:35Z","lastTransitionTime":"2025-11-23T06:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.273430 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.273468 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:45:35 crc kubenswrapper[4559]: E1123 06:45:35.273509 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:45:35 crc kubenswrapper[4559]: E1123 06:45:35.273565 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.358883 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.358939 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.358948 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.358958 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.358965 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:35Z","lastTransitionTime":"2025-11-23T06:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.428950 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/abb6827e-709d-4ea9-8090-e4f541b1ff3f-metrics-certs\") pod \"network-metrics-daemon-svb6z\" (UID: \"abb6827e-709d-4ea9-8090-e4f541b1ff3f\") " pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:45:35 crc kubenswrapper[4559]: E1123 06:45:35.429039 4559 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:45:35 crc kubenswrapper[4559]: E1123 06:45:35.429082 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/abb6827e-709d-4ea9-8090-e4f541b1ff3f-metrics-certs podName:abb6827e-709d-4ea9-8090-e4f541b1ff3f nodeName:}" failed. No retries permitted until 2025-11-23 06:45:37.4290695 +0000 UTC m=+39.451055125 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/abb6827e-709d-4ea9-8090-e4f541b1ff3f-metrics-certs") pod "network-metrics-daemon-svb6z" (UID: "abb6827e-709d-4ea9-8090-e4f541b1ff3f") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.460225 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.460318 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.460387 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.460449 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.460505 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:35Z","lastTransitionTime":"2025-11-23T06:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.561724 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.561755 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.561764 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.561775 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.561782 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:35Z","lastTransitionTime":"2025-11-23T06:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.663157 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.663258 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.663331 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.663389 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.663446 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:35Z","lastTransitionTime":"2025-11-23T06:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.764875 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.764977 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.765047 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.765109 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.765169 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:35Z","lastTransitionTime":"2025-11-23T06:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.866833 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.866858 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.866866 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.866875 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.866883 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:35Z","lastTransitionTime":"2025-11-23T06:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.968946 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.968978 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.968989 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.969002 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:35 crc kubenswrapper[4559]: I1123 06:45:35.969009 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:35Z","lastTransitionTime":"2025-11-23T06:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.070397 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.070425 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.070436 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.070447 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.070455 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:36Z","lastTransitionTime":"2025-11-23T06:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.171939 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.171967 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.171976 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.171986 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.171994 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:36Z","lastTransitionTime":"2025-11-23T06:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.272551 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.272628 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:36 crc kubenswrapper[4559]: E1123 06:45:36.272728 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:45:36 crc kubenswrapper[4559]: E1123 06:45:36.272816 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.273310 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.273337 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.273345 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.273355 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.273361 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:36Z","lastTransitionTime":"2025-11-23T06:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.375385 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.375414 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.375423 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.375434 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.375443 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:36Z","lastTransitionTime":"2025-11-23T06:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.477341 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.477367 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.477375 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.477386 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.477393 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:36Z","lastTransitionTime":"2025-11-23T06:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.579034 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.579068 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.579077 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.579088 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.579097 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:36Z","lastTransitionTime":"2025-11-23T06:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.680670 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.680699 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.680707 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.680717 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.680725 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:36Z","lastTransitionTime":"2025-11-23T06:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.782773 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.782797 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.782806 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.782816 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.782823 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:36Z","lastTransitionTime":"2025-11-23T06:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.884316 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.884350 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.884363 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.884377 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.884390 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:36Z","lastTransitionTime":"2025-11-23T06:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.986275 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.986310 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.986318 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.986330 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:36 crc kubenswrapper[4559]: I1123 06:45:36.986338 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:36Z","lastTransitionTime":"2025-11-23T06:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.087655 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.087681 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.087688 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.087697 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.087705 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:37Z","lastTransitionTime":"2025-11-23T06:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.188816 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.188848 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.188857 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.188867 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.188881 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:37Z","lastTransitionTime":"2025-11-23T06:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.272560 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.272582 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:45:37 crc kubenswrapper[4559]: E1123 06:45:37.272659 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:45:37 crc kubenswrapper[4559]: E1123 06:45:37.272710 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.290487 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.290548 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.290583 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.290618 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.290629 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:37Z","lastTransitionTime":"2025-11-23T06:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.392572 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.392599 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.392607 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.392617 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.392624 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:37Z","lastTransitionTime":"2025-11-23T06:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.444496 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/abb6827e-709d-4ea9-8090-e4f541b1ff3f-metrics-certs\") pod \"network-metrics-daemon-svb6z\" (UID: \"abb6827e-709d-4ea9-8090-e4f541b1ff3f\") " pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:45:37 crc kubenswrapper[4559]: E1123 06:45:37.444672 4559 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:45:37 crc kubenswrapper[4559]: E1123 06:45:37.444712 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/abb6827e-709d-4ea9-8090-e4f541b1ff3f-metrics-certs podName:abb6827e-709d-4ea9-8090-e4f541b1ff3f nodeName:}" failed. No retries permitted until 2025-11-23 06:45:41.444700973 +0000 UTC m=+43.466686588 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/abb6827e-709d-4ea9-8090-e4f541b1ff3f-metrics-certs") pod "network-metrics-daemon-svb6z" (UID: "abb6827e-709d-4ea9-8090-e4f541b1ff3f") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.494296 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.494324 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.494332 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.494360 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.494368 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:37Z","lastTransitionTime":"2025-11-23T06:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.595940 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.595968 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.595976 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.595989 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.595997 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:37Z","lastTransitionTime":"2025-11-23T06:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.697750 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.697780 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.697789 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.697800 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.697810 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:37Z","lastTransitionTime":"2025-11-23T06:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.799223 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.799250 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.799259 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.799269 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.799277 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:37Z","lastTransitionTime":"2025-11-23T06:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.901346 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.901419 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.901429 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.901440 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:37 crc kubenswrapper[4559]: I1123 06:45:37.901448 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:37Z","lastTransitionTime":"2025-11-23T06:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.002986 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.003017 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.003044 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.003067 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.003079 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:38Z","lastTransitionTime":"2025-11-23T06:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.104569 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.104599 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.104608 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.104618 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.104625 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:38Z","lastTransitionTime":"2025-11-23T06:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.206261 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.206290 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.206318 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.206328 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.206336 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:38Z","lastTransitionTime":"2025-11-23T06:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.273172 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.273203 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:38 crc kubenswrapper[4559]: E1123 06:45:38.273253 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:45:38 crc kubenswrapper[4559]: E1123 06:45:38.273379 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.281855 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:38Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.293850 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d455758b0f3b64148d612aca66c657881d20868d079f7d59ddc147936982e6d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d455758b0f3b64148d612aca66c657881d20868d079f7d59ddc147936982e6d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"message\\\":\\\"-machine-approver/machine-approver. OVN-Kubernetes controller took 0.103676017 seconds. No OVN measurement.\\\\nI1123 06:45:29.983899 6024 services_controller.go:445] Built service openshift-kube-apiserver/apiserver LB template configs for network=default: []services.lbConfig(nil)\\\\nI1123 06:45:29.983901 6024 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-s7zkt in node crc\\\\nI1123 06:45:29.983906 6024 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-s7zkt after 0 failed attempt(s)\\\\nI1123 06:45:29.983905 6024 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-rlxnw\\\\nI1123 06:45:29.983910 6024 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-apiserver/apiserver\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.93\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jt56f_openshift-ovn-kubernetes(b12e257e-7887-4795-9221-4db8fd5856dd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:38Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.301663 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:38Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.307436 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.307474 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.307484 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.307497 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.307526 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:38Z","lastTransitionTime":"2025-11-23T06:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.308737 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5bf62d2-3296-4245-970b-2553e647234c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5eb007f4dc7b9bbbed98a8780b4a9ceed7fca2ccd970acd5f7f8a734ae61c55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbebf00280a1ffac1595cd96655356e8567fe98e96ab346e0225f2fcc1c4de9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9k8xh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:38Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.316843 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:38Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.323434 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:38Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.336479 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:38Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.343110 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:38Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.351240 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:38Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.359459 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:38Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.367775 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:38Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.374785 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:38Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.383142 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:38Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.393360 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56e33b4cd5df561efe04dcbbf6688c9edca01209b83ca44b40fc61cf562b2e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:38Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.400371 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svb6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abb6827e-709d-4ea9-8090-e4f541b1ff3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svb6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:38Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.408918 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:38Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.409714 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.409741 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.409749 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.409760 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.409788 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:38Z","lastTransitionTime":"2025-11-23T06:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.511272 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.511299 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.511307 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.511316 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.511324 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:38Z","lastTransitionTime":"2025-11-23T06:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.613040 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.613073 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.613081 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.613094 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.613102 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:38Z","lastTransitionTime":"2025-11-23T06:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.714958 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.715050 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.715062 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.715072 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.715079 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:38Z","lastTransitionTime":"2025-11-23T06:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.816366 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.816391 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.816415 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.816426 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.816432 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:38Z","lastTransitionTime":"2025-11-23T06:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.918386 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.918412 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.918421 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.918439 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:38 crc kubenswrapper[4559]: I1123 06:45:38.918448 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:38Z","lastTransitionTime":"2025-11-23T06:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.020009 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.020034 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.020041 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.020050 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.020058 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:39Z","lastTransitionTime":"2025-11-23T06:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.090673 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.090711 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.090721 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.090732 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.090741 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:39Z","lastTransitionTime":"2025-11-23T06:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:39 crc kubenswrapper[4559]: E1123 06:45:39.099323 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:39Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.101411 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.101440 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.101449 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.101513 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.101522 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:39Z","lastTransitionTime":"2025-11-23T06:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:39 crc kubenswrapper[4559]: E1123 06:45:39.109043 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:39Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.110959 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.110983 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.110991 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.111000 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.111007 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:39Z","lastTransitionTime":"2025-11-23T06:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:39 crc kubenswrapper[4559]: E1123 06:45:39.118368 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:39Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.120236 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.120261 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.120269 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.120278 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.120285 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:39Z","lastTransitionTime":"2025-11-23T06:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:39 crc kubenswrapper[4559]: E1123 06:45:39.127512 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:39Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.129286 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.129305 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.129314 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.129322 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.129329 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:39Z","lastTransitionTime":"2025-11-23T06:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:39 crc kubenswrapper[4559]: E1123 06:45:39.136848 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:39Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:39 crc kubenswrapper[4559]: E1123 06:45:39.136945 4559 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.137718 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.137741 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.137749 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.137757 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.137765 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:39Z","lastTransitionTime":"2025-11-23T06:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.239548 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.239571 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.239578 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.239588 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.239595 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:39Z","lastTransitionTime":"2025-11-23T06:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.273137 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:39 crc kubenswrapper[4559]: E1123 06:45:39.273220 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.273139 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:45:39 crc kubenswrapper[4559]: E1123 06:45:39.273313 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.341510 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.341608 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.341636 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.341683 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.341691 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:39Z","lastTransitionTime":"2025-11-23T06:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.443692 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.443720 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.443728 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.443741 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.443749 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:39Z","lastTransitionTime":"2025-11-23T06:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.545409 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.545444 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.545454 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.545467 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.545478 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:39Z","lastTransitionTime":"2025-11-23T06:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.647282 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.647319 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.647329 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.647341 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.647350 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:39Z","lastTransitionTime":"2025-11-23T06:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.749061 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.749087 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.749095 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.749106 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.749113 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:39Z","lastTransitionTime":"2025-11-23T06:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.850546 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.850577 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.850587 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.850600 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.850609 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:39Z","lastTransitionTime":"2025-11-23T06:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.952312 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.952340 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.952347 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.952356 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:39 crc kubenswrapper[4559]: I1123 06:45:39.952365 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:39Z","lastTransitionTime":"2025-11-23T06:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.053984 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.054027 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.054035 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.054045 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.054052 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:40Z","lastTransitionTime":"2025-11-23T06:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.155336 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.155360 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.155370 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.155380 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.155387 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:40Z","lastTransitionTime":"2025-11-23T06:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.257204 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.257233 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.257244 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.257256 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.257266 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:40Z","lastTransitionTime":"2025-11-23T06:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.273191 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.273248 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:40 crc kubenswrapper[4559]: E1123 06:45:40.273338 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:45:40 crc kubenswrapper[4559]: E1123 06:45:40.273395 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.358539 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.358565 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.358573 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.358584 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.358594 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:40Z","lastTransitionTime":"2025-11-23T06:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.460356 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.460381 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.460388 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.460399 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.460406 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:40Z","lastTransitionTime":"2025-11-23T06:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.562146 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.562171 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.562178 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.562187 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.562194 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:40Z","lastTransitionTime":"2025-11-23T06:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.663853 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.663928 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.663939 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.663949 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.663958 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:40Z","lastTransitionTime":"2025-11-23T06:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.765994 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.766048 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.766057 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.766066 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.766073 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:40Z","lastTransitionTime":"2025-11-23T06:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.867590 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.867615 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.867623 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.867632 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.867659 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:40Z","lastTransitionTime":"2025-11-23T06:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.969467 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.969532 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.969543 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.969554 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:40 crc kubenswrapper[4559]: I1123 06:45:40.969562 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:40Z","lastTransitionTime":"2025-11-23T06:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.070602 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.070627 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.070634 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.070658 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.070667 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:41Z","lastTransitionTime":"2025-11-23T06:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.172521 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.172563 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.172572 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.172584 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.172592 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:41Z","lastTransitionTime":"2025-11-23T06:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.272718 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.272744 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:45:41 crc kubenswrapper[4559]: E1123 06:45:41.272804 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:45:41 crc kubenswrapper[4559]: E1123 06:45:41.272863 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.273408 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.273434 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.273443 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.273455 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.273462 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:41Z","lastTransitionTime":"2025-11-23T06:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.375193 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.375217 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.375224 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.375234 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.375241 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:41Z","lastTransitionTime":"2025-11-23T06:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.473637 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/abb6827e-709d-4ea9-8090-e4f541b1ff3f-metrics-certs\") pod \"network-metrics-daemon-svb6z\" (UID: \"abb6827e-709d-4ea9-8090-e4f541b1ff3f\") " pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:45:41 crc kubenswrapper[4559]: E1123 06:45:41.473775 4559 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:45:41 crc kubenswrapper[4559]: E1123 06:45:41.473846 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/abb6827e-709d-4ea9-8090-e4f541b1ff3f-metrics-certs podName:abb6827e-709d-4ea9-8090-e4f541b1ff3f nodeName:}" failed. No retries permitted until 2025-11-23 06:45:49.473831065 +0000 UTC m=+51.495816678 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/abb6827e-709d-4ea9-8090-e4f541b1ff3f-metrics-certs") pod "network-metrics-daemon-svb6z" (UID: "abb6827e-709d-4ea9-8090-e4f541b1ff3f") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.476686 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.476719 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.476728 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.476739 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.476747 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:41Z","lastTransitionTime":"2025-11-23T06:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.578729 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.578762 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.578772 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.578784 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.578793 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:41Z","lastTransitionTime":"2025-11-23T06:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.680326 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.680346 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.680354 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.680366 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.680373 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:41Z","lastTransitionTime":"2025-11-23T06:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.782063 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.782090 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.782099 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.782108 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.782116 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:41Z","lastTransitionTime":"2025-11-23T06:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.883682 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.883707 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.883715 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.883726 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.883733 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:41Z","lastTransitionTime":"2025-11-23T06:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.984770 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.984793 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.984801 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.984810 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:41 crc kubenswrapper[4559]: I1123 06:45:41.984817 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:41Z","lastTransitionTime":"2025-11-23T06:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.086887 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.086912 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.086920 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.086930 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.086937 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:42Z","lastTransitionTime":"2025-11-23T06:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.191780 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.191817 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.191826 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.191859 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.191874 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:42Z","lastTransitionTime":"2025-11-23T06:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.272761 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.272769 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:42 crc kubenswrapper[4559]: E1123 06:45:42.272966 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:45:42 crc kubenswrapper[4559]: E1123 06:45:42.272879 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.293885 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.293926 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.293935 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.293944 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.293951 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:42Z","lastTransitionTime":"2025-11-23T06:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.395736 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.395762 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.395770 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.395779 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.395802 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:42Z","lastTransitionTime":"2025-11-23T06:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.496773 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.496804 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.496814 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.496830 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.496839 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:42Z","lastTransitionTime":"2025-11-23T06:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.597731 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.597759 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.597770 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.597781 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.597790 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:42Z","lastTransitionTime":"2025-11-23T06:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.699683 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.699781 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.699796 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.699825 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.699832 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:42Z","lastTransitionTime":"2025-11-23T06:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.801193 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.801227 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.801237 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.801248 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.801258 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:42Z","lastTransitionTime":"2025-11-23T06:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.902980 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.903012 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.903020 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.903032 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:42 crc kubenswrapper[4559]: I1123 06:45:42.903040 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:42Z","lastTransitionTime":"2025-11-23T06:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.004494 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.004516 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.004524 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.004533 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.004541 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:43Z","lastTransitionTime":"2025-11-23T06:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.106514 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.106534 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.106542 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.106551 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.106558 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:43Z","lastTransitionTime":"2025-11-23T06:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.207916 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.207951 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.207961 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.207973 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.207982 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:43Z","lastTransitionTime":"2025-11-23T06:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.272605 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.272665 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:43 crc kubenswrapper[4559]: E1123 06:45:43.272699 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:45:43 crc kubenswrapper[4559]: E1123 06:45:43.272937 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.273207 4559 scope.go:117] "RemoveContainer" containerID="d455758b0f3b64148d612aca66c657881d20868d079f7d59ddc147936982e6d3" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.309251 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.309278 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.309286 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.309297 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.309321 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:43Z","lastTransitionTime":"2025-11-23T06:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.411049 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.411079 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.411087 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.411101 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.411109 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:43Z","lastTransitionTime":"2025-11-23T06:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.440383 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jt56f_b12e257e-7887-4795-9221-4db8fd5856dd/ovnkube-controller/1.log" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.442531 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" event={"ID":"b12e257e-7887-4795-9221-4db8fd5856dd","Type":"ContainerStarted","Data":"f0a58330faa93c0acb412514c8429ae62f58ef1a847a66a9a3392736f5d8d88d"} Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.442634 4559 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.452816 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:43Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.460912 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:43Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.469916 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:43Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.479719 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:43Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.490498 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:43Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.502009 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:43Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.510782 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:43Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.512823 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.512859 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.512869 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.512888 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.512897 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:43Z","lastTransitionTime":"2025-11-23T06:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.521789 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:43Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.535490 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56e33b4cd5df561efe04dcbbf6688c9edca01209b83ca44b40fc61cf562b2e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:43Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.545626 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svb6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abb6827e-709d-4ea9-8090-e4f541b1ff3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svb6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:43Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.555526 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:43Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.564245 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:43Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.576628 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0a58330faa93c0acb412514c8429ae62f58ef1a847a66a9a3392736f5d8d88d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d455758b0f3b64148d612aca66c657881d20868d079f7d59ddc147936982e6d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"message\\\":\\\"-machine-approver/machine-approver. OVN-Kubernetes controller took 0.103676017 seconds. No OVN measurement.\\\\nI1123 06:45:29.983899 6024 services_controller.go:445] Built service openshift-kube-apiserver/apiserver LB template configs for network=default: []services.lbConfig(nil)\\\\nI1123 06:45:29.983901 6024 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-s7zkt in node crc\\\\nI1123 06:45:29.983906 6024 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-s7zkt after 0 failed attempt(s)\\\\nI1123 06:45:29.983905 6024 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-rlxnw\\\\nI1123 06:45:29.983910 6024 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-apiserver/apiserver\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.93\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:43Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.584722 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:43Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.591096 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:43Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.601254 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5bf62d2-3296-4245-970b-2553e647234c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5eb007f4dc7b9bbbed98a8780b4a9ceed7fca2ccd970acd5f7f8a734ae61c55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbebf00280a1ffac1595cd96655356e8567fe98e96ab346e0225f2fcc1c4de9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9k8xh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:43Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.614463 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.614495 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.614505 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.614517 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.614526 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:43Z","lastTransitionTime":"2025-11-23T06:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.716580 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.716622 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.716630 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.716677 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.716688 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:43Z","lastTransitionTime":"2025-11-23T06:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.818739 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.818769 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.818777 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.818792 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.818800 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:43Z","lastTransitionTime":"2025-11-23T06:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.921102 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.921138 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.921147 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.921159 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:43 crc kubenswrapper[4559]: I1123 06:45:43.921167 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:43Z","lastTransitionTime":"2025-11-23T06:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.023321 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.023360 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.023368 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.023383 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.023391 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:44Z","lastTransitionTime":"2025-11-23T06:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.124736 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.124767 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.124778 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.124789 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.124797 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:44Z","lastTransitionTime":"2025-11-23T06:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.226745 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.226782 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.226792 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.226806 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.226817 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:44Z","lastTransitionTime":"2025-11-23T06:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.272752 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.272808 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:44 crc kubenswrapper[4559]: E1123 06:45:44.272874 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:45:44 crc kubenswrapper[4559]: E1123 06:45:44.272940 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.328118 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.328147 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.328156 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.328168 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.328177 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:44Z","lastTransitionTime":"2025-11-23T06:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.430033 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.430072 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.430081 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.430093 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.430102 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:44Z","lastTransitionTime":"2025-11-23T06:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.446118 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jt56f_b12e257e-7887-4795-9221-4db8fd5856dd/ovnkube-controller/2.log" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.446549 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jt56f_b12e257e-7887-4795-9221-4db8fd5856dd/ovnkube-controller/1.log" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.448161 4559 generic.go:334] "Generic (PLEG): container finished" podID="b12e257e-7887-4795-9221-4db8fd5856dd" containerID="f0a58330faa93c0acb412514c8429ae62f58ef1a847a66a9a3392736f5d8d88d" exitCode=1 Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.448187 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" event={"ID":"b12e257e-7887-4795-9221-4db8fd5856dd","Type":"ContainerDied","Data":"f0a58330faa93c0acb412514c8429ae62f58ef1a847a66a9a3392736f5d8d88d"} Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.448209 4559 scope.go:117] "RemoveContainer" containerID="d455758b0f3b64148d612aca66c657881d20868d079f7d59ddc147936982e6d3" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.448695 4559 scope.go:117] "RemoveContainer" containerID="f0a58330faa93c0acb412514c8429ae62f58ef1a847a66a9a3392736f5d8d88d" Nov 23 06:45:44 crc kubenswrapper[4559]: E1123 06:45:44.448807 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jt56f_openshift-ovn-kubernetes(b12e257e-7887-4795-9221-4db8fd5856dd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.460383 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56e33b4cd5df561efe04dcbbf6688c9edca01209b83ca44b40fc61cf562b2e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:44Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.468628 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svb6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abb6827e-709d-4ea9-8090-e4f541b1ff3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svb6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:44Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.476995 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:44Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.484906 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:44Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.493493 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:44Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.500473 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:44Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.508108 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:44Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.515511 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:44Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.522734 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:44Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.531825 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.531849 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.531857 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.531868 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.531876 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:44Z","lastTransitionTime":"2025-11-23T06:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.534886 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0a58330faa93c0acb412514c8429ae62f58ef1a847a66a9a3392736f5d8d88d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d455758b0f3b64148d612aca66c657881d20868d079f7d59ddc147936982e6d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"message\\\":\\\"-machine-approver/machine-approver. OVN-Kubernetes controller took 0.103676017 seconds. No OVN measurement.\\\\nI1123 06:45:29.983899 6024 services_controller.go:445] Built service openshift-kube-apiserver/apiserver LB template configs for network=default: []services.lbConfig(nil)\\\\nI1123 06:45:29.983901 6024 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-s7zkt in node crc\\\\nI1123 06:45:29.983906 6024 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-s7zkt after 0 failed attempt(s)\\\\nI1123 06:45:29.983905 6024 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-rlxnw\\\\nI1123 06:45:29.983910 6024 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-apiserver/apiserver\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.93\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0a58330faa93c0acb412514c8429ae62f58ef1a847a66a9a3392736f5d8d88d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:45:43Z\\\",\\\"message\\\":\\\"27.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:43Z is after 2025-08-24T17:21:41Z]\\\\nI1123 06:45:43.858842 6266 services_controller.go:434] Service openshift-operator-lifecycle-manager/catalog-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{catalog-operator-metrics openshift-operator-lifecycle-manager a8e45f62-9ba9-441e-8ed9-b3756498d7e7 4615 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:catalog-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:catalog-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0076825e7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: catalog-operator,},ClusterIP:10.217\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:44Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.542751 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:44Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.549193 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:44Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.556659 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5bf62d2-3296-4245-970b-2553e647234c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5eb007f4dc7b9bbbed98a8780b4a9ceed7fca2ccd970acd5f7f8a734ae61c55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbebf00280a1ffac1595cd96655356e8567fe98e96ab346e0225f2fcc1c4de9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9k8xh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:44Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.564483 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:44Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.571920 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:44Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.577870 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:44Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.633382 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.633405 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.633414 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.633426 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.633446 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:44Z","lastTransitionTime":"2025-11-23T06:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.735683 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.735711 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.735720 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.735730 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.735737 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:44Z","lastTransitionTime":"2025-11-23T06:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.839319 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.839340 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.839349 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.839358 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.839366 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:44Z","lastTransitionTime":"2025-11-23T06:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.941506 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.941550 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.941559 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.941571 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:44 crc kubenswrapper[4559]: I1123 06:45:44.941579 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:44Z","lastTransitionTime":"2025-11-23T06:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.043458 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.043487 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.043497 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.043510 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.043518 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:45Z","lastTransitionTime":"2025-11-23T06:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.145325 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.145360 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.145370 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.145382 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.145391 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:45Z","lastTransitionTime":"2025-11-23T06:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.246617 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.246656 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.246665 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.246677 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.246684 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:45Z","lastTransitionTime":"2025-11-23T06:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.273271 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.273336 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:45:45 crc kubenswrapper[4559]: E1123 06:45:45.273355 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:45:45 crc kubenswrapper[4559]: E1123 06:45:45.273498 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.348911 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.348940 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.348949 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.348960 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.348968 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:45Z","lastTransitionTime":"2025-11-23T06:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.449995 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.450021 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.450029 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.450040 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.450047 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:45Z","lastTransitionTime":"2025-11-23T06:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.451279 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jt56f_b12e257e-7887-4795-9221-4db8fd5856dd/ovnkube-controller/2.log" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.551618 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.551657 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.551666 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.551676 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.551683 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:45Z","lastTransitionTime":"2025-11-23T06:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.653472 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.653504 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.653512 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.653524 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.653533 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:45Z","lastTransitionTime":"2025-11-23T06:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.754799 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.754833 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.754843 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.754854 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.754861 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:45Z","lastTransitionTime":"2025-11-23T06:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.856115 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.856149 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.856159 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.856171 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.856181 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:45Z","lastTransitionTime":"2025-11-23T06:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.957493 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.957524 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.957534 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.957544 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:45 crc kubenswrapper[4559]: I1123 06:45:45.957551 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:45Z","lastTransitionTime":"2025-11-23T06:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.059204 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.059238 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.059249 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.059261 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.059270 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:46Z","lastTransitionTime":"2025-11-23T06:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.161188 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.161217 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.161226 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.161237 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.161249 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:46Z","lastTransitionTime":"2025-11-23T06:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.262357 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.262385 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.262393 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.262402 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.262424 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:46Z","lastTransitionTime":"2025-11-23T06:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.272849 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:46 crc kubenswrapper[4559]: E1123 06:45:46.272925 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.272856 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:46 crc kubenswrapper[4559]: E1123 06:45:46.273008 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.364163 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.364195 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.364204 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.364216 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.364224 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:46Z","lastTransitionTime":"2025-11-23T06:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.466137 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.466186 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.466196 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.466210 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.466222 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:46Z","lastTransitionTime":"2025-11-23T06:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.567895 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.567918 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.567927 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.567936 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.567944 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:46Z","lastTransitionTime":"2025-11-23T06:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.669998 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.670023 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.670031 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.670040 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.670047 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:46Z","lastTransitionTime":"2025-11-23T06:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.772085 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.772118 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.772127 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.772142 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.772150 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:46Z","lastTransitionTime":"2025-11-23T06:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.873198 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.873256 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.873266 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.873279 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.873287 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:46Z","lastTransitionTime":"2025-11-23T06:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.974767 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.974818 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.974832 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.974849 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:46 crc kubenswrapper[4559]: I1123 06:45:46.974862 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:46Z","lastTransitionTime":"2025-11-23T06:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.077092 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.077144 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.077154 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.077165 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.077173 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:47Z","lastTransitionTime":"2025-11-23T06:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.179155 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.179185 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.179194 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.179204 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.179210 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:47Z","lastTransitionTime":"2025-11-23T06:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.273408 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.273441 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:45:47 crc kubenswrapper[4559]: E1123 06:45:47.273493 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:45:47 crc kubenswrapper[4559]: E1123 06:45:47.273551 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.280175 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.280200 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.280210 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.280219 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.280226 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:47Z","lastTransitionTime":"2025-11-23T06:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.381830 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.381862 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.381873 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.381883 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.381893 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:47Z","lastTransitionTime":"2025-11-23T06:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.483860 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.483881 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.483888 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.483897 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.483905 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:47Z","lastTransitionTime":"2025-11-23T06:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.585209 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.585251 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.585260 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.585268 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.585276 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:47Z","lastTransitionTime":"2025-11-23T06:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.686904 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.686940 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.686953 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.686967 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.686977 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:47Z","lastTransitionTime":"2025-11-23T06:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.788795 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.788817 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.788825 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.788834 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.788841 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:47Z","lastTransitionTime":"2025-11-23T06:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.889919 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.889944 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.889953 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.889963 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.889971 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:47Z","lastTransitionTime":"2025-11-23T06:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.925542 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:47 crc kubenswrapper[4559]: E1123 06:45:47.925672 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:45:47 crc kubenswrapper[4559]: E1123 06:45:47.925687 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:45:47 crc kubenswrapper[4559]: E1123 06:45:47.925696 4559 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:45:47 crc kubenswrapper[4559]: E1123 06:45:47.925733 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-23 06:46:19.925721706 +0000 UTC m=+81.947707320 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.991669 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.991695 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.991703 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.991712 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:47 crc kubenswrapper[4559]: I1123 06:45:47.991719 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:47Z","lastTransitionTime":"2025-11-23T06:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.026057 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.026111 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:48 crc kubenswrapper[4559]: E1123 06:45:48.026214 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:46:20.02619477 +0000 UTC m=+82.048180384 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:45:48 crc kubenswrapper[4559]: E1123 06:45:48.026222 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:45:48 crc kubenswrapper[4559]: E1123 06:45:48.026238 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:45:48 crc kubenswrapper[4559]: E1123 06:45:48.026246 4559 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:45:48 crc kubenswrapper[4559]: E1123 06:45:48.026269 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-23 06:46:20.026261986 +0000 UTC m=+82.048247590 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.093587 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.093613 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.093621 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.093631 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.093639 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:48Z","lastTransitionTime":"2025-11-23T06:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.127352 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.127412 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:48 crc kubenswrapper[4559]: E1123 06:45:48.127474 4559 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:45:48 crc kubenswrapper[4559]: E1123 06:45:48.127522 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:46:20.127512733 +0000 UTC m=+82.149498346 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:45:48 crc kubenswrapper[4559]: E1123 06:45:48.127479 4559 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:45:48 crc kubenswrapper[4559]: E1123 06:45:48.127611 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:46:20.127602059 +0000 UTC m=+82.149587673 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.195236 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.195257 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.195265 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.195274 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.195280 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:48Z","lastTransitionTime":"2025-11-23T06:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.254222 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.254744 4559 scope.go:117] "RemoveContainer" containerID="f0a58330faa93c0acb412514c8429ae62f58ef1a847a66a9a3392736f5d8d88d" Nov 23 06:45:48 crc kubenswrapper[4559]: E1123 06:45:48.254875 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jt56f_openshift-ovn-kubernetes(b12e257e-7887-4795-9221-4db8fd5856dd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.263559 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.271082 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.272789 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:48 crc kubenswrapper[4559]: E1123 06:45:48.272872 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.272800 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:48 crc kubenswrapper[4559]: E1123 06:45:48.272992 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.279361 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.287860 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56e33b4cd5df561efe04dcbbf6688c9edca01209b83ca44b40fc61cf562b2e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.293832 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svb6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abb6827e-709d-4ea9-8090-e4f541b1ff3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svb6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.296431 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.296474 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.296483 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.296495 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.296504 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:48Z","lastTransitionTime":"2025-11-23T06:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.302076 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.309837 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.321604 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0a58330faa93c0acb412514c8429ae62f58ef1a847a66a9a3392736f5d8d88d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0a58330faa93c0acb412514c8429ae62f58ef1a847a66a9a3392736f5d8d88d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:45:43Z\\\",\\\"message\\\":\\\"27.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:43Z is after 2025-08-24T17:21:41Z]\\\\nI1123 06:45:43.858842 6266 services_controller.go:434] Service openshift-operator-lifecycle-manager/catalog-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{catalog-operator-metrics openshift-operator-lifecycle-manager a8e45f62-9ba9-441e-8ed9-b3756498d7e7 4615 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:catalog-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:catalog-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0076825e7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: catalog-operator,},ClusterIP:10.217\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jt56f_openshift-ovn-kubernetes(b12e257e-7887-4795-9221-4db8fd5856dd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.333993 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.341536 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.349094 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.355200 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.362555 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5bf62d2-3296-4245-970b-2553e647234c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5eb007f4dc7b9bbbed98a8780b4a9ceed7fca2ccd970acd5f7f8a734ae61c55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbebf00280a1ffac1595cd96655356e8567fe98e96ab346e0225f2fcc1c4de9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9k8xh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.369069 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.376852 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.384162 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.391250 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.398059 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.398127 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.398148 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.398156 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.398167 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.398175 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:48Z","lastTransitionTime":"2025-11-23T06:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.410483 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0a58330faa93c0acb412514c8429ae62f58ef1a847a66a9a3392736f5d8d88d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0a58330faa93c0acb412514c8429ae62f58ef1a847a66a9a3392736f5d8d88d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:45:43Z\\\",\\\"message\\\":\\\"27.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:43Z is after 2025-08-24T17:21:41Z]\\\\nI1123 06:45:43.858842 6266 services_controller.go:434] Service openshift-operator-lifecycle-manager/catalog-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{catalog-operator-metrics openshift-operator-lifecycle-manager a8e45f62-9ba9-441e-8ed9-b3756498d7e7 4615 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:catalog-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:catalog-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0076825e7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: catalog-operator,},ClusterIP:10.217\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jt56f_openshift-ovn-kubernetes(b12e257e-7887-4795-9221-4db8fd5856dd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.418169 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.426357 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.433801 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5bf62d2-3296-4245-970b-2553e647234c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5eb007f4dc7b9bbbed98a8780b4a9ceed7fca2ccd970acd5f7f8a734ae61c55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbebf00280a1ffac1595cd96655356e8567fe98e96ab346e0225f2fcc1c4de9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9k8xh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.441726 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.448976 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.455014 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.463582 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.472815 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56e33b4cd5df561efe04dcbbf6688c9edca01209b83ca44b40fc61cf562b2e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.479242 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svb6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abb6827e-709d-4ea9-8090-e4f541b1ff3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svb6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.487219 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.494602 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.499459 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.499486 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.499495 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.499506 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.499514 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:48Z","lastTransitionTime":"2025-11-23T06:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.502339 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.509485 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:48Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.601118 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.601143 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.601152 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.601164 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.601173 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:48Z","lastTransitionTime":"2025-11-23T06:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.702257 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.702283 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.702292 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.702303 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.702310 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:48Z","lastTransitionTime":"2025-11-23T06:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.803631 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.803679 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.803688 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.803700 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.803708 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:48Z","lastTransitionTime":"2025-11-23T06:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.905717 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.905745 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.905753 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.905764 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:48 crc kubenswrapper[4559]: I1123 06:45:48.905771 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:48Z","lastTransitionTime":"2025-11-23T06:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.007672 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.007701 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.007710 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.007722 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.007731 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:49Z","lastTransitionTime":"2025-11-23T06:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.109307 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.109333 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.109342 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.109352 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.109360 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:49Z","lastTransitionTime":"2025-11-23T06:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.210501 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.210532 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.210540 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.210552 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.210560 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:49Z","lastTransitionTime":"2025-11-23T06:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.272941 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.272975 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:45:49 crc kubenswrapper[4559]: E1123 06:45:49.273020 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:45:49 crc kubenswrapper[4559]: E1123 06:45:49.273078 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.312319 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.312352 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.312362 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.312382 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.312391 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:49Z","lastTransitionTime":"2025-11-23T06:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.406439 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.406465 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.406473 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.406483 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.406490 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:49Z","lastTransitionTime":"2025-11-23T06:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:49 crc kubenswrapper[4559]: E1123 06:45:49.414906 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:49Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.417165 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.417190 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.417199 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.417208 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.417215 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:49Z","lastTransitionTime":"2025-11-23T06:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:49 crc kubenswrapper[4559]: E1123 06:45:49.425534 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:49Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.427713 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.427736 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.427744 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.427755 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.427764 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:49Z","lastTransitionTime":"2025-11-23T06:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:49 crc kubenswrapper[4559]: E1123 06:45:49.435512 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:49Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.437707 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.437733 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.437742 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.437753 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.437760 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:49Z","lastTransitionTime":"2025-11-23T06:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:49 crc kubenswrapper[4559]: E1123 06:45:49.445126 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:49Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.447175 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.447204 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.447212 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.447225 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.447231 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:49Z","lastTransitionTime":"2025-11-23T06:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:49 crc kubenswrapper[4559]: E1123 06:45:49.454380 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:49Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:49 crc kubenswrapper[4559]: E1123 06:45:49.454487 4559 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.455339 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.455412 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.455422 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.455432 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.455439 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:49Z","lastTransitionTime":"2025-11-23T06:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.537095 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/abb6827e-709d-4ea9-8090-e4f541b1ff3f-metrics-certs\") pod \"network-metrics-daemon-svb6z\" (UID: \"abb6827e-709d-4ea9-8090-e4f541b1ff3f\") " pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:45:49 crc kubenswrapper[4559]: E1123 06:45:49.537187 4559 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:45:49 crc kubenswrapper[4559]: E1123 06:45:49.537244 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/abb6827e-709d-4ea9-8090-e4f541b1ff3f-metrics-certs podName:abb6827e-709d-4ea9-8090-e4f541b1ff3f nodeName:}" failed. No retries permitted until 2025-11-23 06:46:05.537230838 +0000 UTC m=+67.559216462 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/abb6827e-709d-4ea9-8090-e4f541b1ff3f-metrics-certs") pod "network-metrics-daemon-svb6z" (UID: "abb6827e-709d-4ea9-8090-e4f541b1ff3f") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.556934 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.556964 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.556974 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.556984 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.556993 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:49Z","lastTransitionTime":"2025-11-23T06:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.658995 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.659020 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.659028 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.659039 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.659046 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:49Z","lastTransitionTime":"2025-11-23T06:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.761072 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.761103 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.761115 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.761126 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.761134 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:49Z","lastTransitionTime":"2025-11-23T06:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.862526 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.862551 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.862560 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.862569 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.862577 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:49Z","lastTransitionTime":"2025-11-23T06:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.964193 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.964212 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.964220 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.964231 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:49 crc kubenswrapper[4559]: I1123 06:45:49.964237 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:49Z","lastTransitionTime":"2025-11-23T06:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.066851 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.066877 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.066885 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.066895 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.066902 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:50Z","lastTransitionTime":"2025-11-23T06:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.169135 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.169250 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.169309 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.169400 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.169468 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:50Z","lastTransitionTime":"2025-11-23T06:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.271838 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.271867 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.271876 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.271887 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.271894 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:50Z","lastTransitionTime":"2025-11-23T06:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.273663 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:50 crc kubenswrapper[4559]: E1123 06:45:50.273761 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.273799 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:50 crc kubenswrapper[4559]: E1123 06:45:50.273887 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.373328 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.373373 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.373383 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.373394 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.373404 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:50Z","lastTransitionTime":"2025-11-23T06:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.475281 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.475315 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.475326 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.475339 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.475350 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:50Z","lastTransitionTime":"2025-11-23T06:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.577265 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.577293 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.577301 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.577312 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.577320 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:50Z","lastTransitionTime":"2025-11-23T06:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.678858 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.678916 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.678929 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.678948 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.678961 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:50Z","lastTransitionTime":"2025-11-23T06:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.781381 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.781416 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.781425 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.781441 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.781451 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:50Z","lastTransitionTime":"2025-11-23T06:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.883266 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.883311 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.883319 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.883337 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.883349 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:50Z","lastTransitionTime":"2025-11-23T06:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.984542 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.984602 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.984615 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.984633 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:50 crc kubenswrapper[4559]: I1123 06:45:50.984666 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:50Z","lastTransitionTime":"2025-11-23T06:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.086697 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.086724 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.086733 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.086759 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.086769 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:51Z","lastTransitionTime":"2025-11-23T06:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.188844 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.188892 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.188902 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.188919 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.188933 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:51Z","lastTransitionTime":"2025-11-23T06:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.273534 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.273543 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:45:51 crc kubenswrapper[4559]: E1123 06:45:51.273767 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:45:51 crc kubenswrapper[4559]: E1123 06:45:51.273634 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.290741 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.290787 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.290799 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.290815 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.290827 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:51Z","lastTransitionTime":"2025-11-23T06:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.392623 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.392680 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.392690 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.392701 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.392709 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:51Z","lastTransitionTime":"2025-11-23T06:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.494703 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.494746 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.494757 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.494774 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.494785 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:51Z","lastTransitionTime":"2025-11-23T06:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.596669 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.596691 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.596700 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.596712 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.596719 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:51Z","lastTransitionTime":"2025-11-23T06:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.699685 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.699736 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.699755 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.699773 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.699786 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:51Z","lastTransitionTime":"2025-11-23T06:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.802469 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.802520 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.802531 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.802551 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.802562 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:51Z","lastTransitionTime":"2025-11-23T06:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.904775 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.904816 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.904831 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.904851 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:51 crc kubenswrapper[4559]: I1123 06:45:51.904861 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:51Z","lastTransitionTime":"2025-11-23T06:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.006815 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.006849 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.006874 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.006888 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.006899 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:52Z","lastTransitionTime":"2025-11-23T06:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.108908 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.108988 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.109004 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.109030 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.109044 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:52Z","lastTransitionTime":"2025-11-23T06:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.211819 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.211853 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.211863 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.211881 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.211892 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:52Z","lastTransitionTime":"2025-11-23T06:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.272889 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.272895 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:52 crc kubenswrapper[4559]: E1123 06:45:52.273007 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:45:52 crc kubenswrapper[4559]: E1123 06:45:52.273158 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.314479 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.314511 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.314521 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.314532 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.314540 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:52Z","lastTransitionTime":"2025-11-23T06:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.416757 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.416789 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.416799 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.416809 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.416818 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:52Z","lastTransitionTime":"2025-11-23T06:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.518703 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.518725 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.518735 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.518747 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.518754 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:52Z","lastTransitionTime":"2025-11-23T06:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.621265 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.621351 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.621367 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.621395 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.621416 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:52Z","lastTransitionTime":"2025-11-23T06:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.723372 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.723407 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.723417 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.723431 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.723454 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:52Z","lastTransitionTime":"2025-11-23T06:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.825224 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.825288 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.825299 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.825317 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.825341 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:52Z","lastTransitionTime":"2025-11-23T06:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.927974 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.928009 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.928021 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.928035 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:52 crc kubenswrapper[4559]: I1123 06:45:52.928044 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:52Z","lastTransitionTime":"2025-11-23T06:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.029812 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.030031 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.030040 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.030051 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.030059 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:53Z","lastTransitionTime":"2025-11-23T06:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.132092 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.132122 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.132131 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.132144 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.132151 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:53Z","lastTransitionTime":"2025-11-23T06:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.234028 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.234072 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.234081 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.234098 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.234115 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:53Z","lastTransitionTime":"2025-11-23T06:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.272501 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.272538 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:45:53 crc kubenswrapper[4559]: E1123 06:45:53.272596 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:45:53 crc kubenswrapper[4559]: E1123 06:45:53.272667 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.314568 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.321358 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.324887 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.332959 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.335707 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.335738 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.335748 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.335762 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.335773 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:53Z","lastTransitionTime":"2025-11-23T06:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.340136 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.348733 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.357462 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.365838 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.373020 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.381317 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.390208 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56e33b4cd5df561efe04dcbbf6688c9edca01209b83ca44b40fc61cf562b2e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.396756 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svb6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abb6827e-709d-4ea9-8090-e4f541b1ff3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svb6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.404288 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.411153 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.422615 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0a58330faa93c0acb412514c8429ae62f58ef1a847a66a9a3392736f5d8d88d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0a58330faa93c0acb412514c8429ae62f58ef1a847a66a9a3392736f5d8d88d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:45:43Z\\\",\\\"message\\\":\\\"27.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:43Z is after 2025-08-24T17:21:41Z]\\\\nI1123 06:45:43.858842 6266 services_controller.go:434] Service openshift-operator-lifecycle-manager/catalog-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{catalog-operator-metrics openshift-operator-lifecycle-manager a8e45f62-9ba9-441e-8ed9-b3756498d7e7 4615 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:catalog-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:catalog-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0076825e7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: catalog-operator,},ClusterIP:10.217\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jt56f_openshift-ovn-kubernetes(b12e257e-7887-4795-9221-4db8fd5856dd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.429134 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.436340 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5bf62d2-3296-4245-970b-2553e647234c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5eb007f4dc7b9bbbed98a8780b4a9ceed7fca2ccd970acd5f7f8a734ae61c55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbebf00280a1ffac1595cd96655356e8567fe98e96ab346e0225f2fcc1c4de9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9k8xh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.437812 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.437846 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.437855 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.437870 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.437880 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:53Z","lastTransitionTime":"2025-11-23T06:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.444877 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:53Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.539380 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.539407 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.539420 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.539432 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.539440 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:53Z","lastTransitionTime":"2025-11-23T06:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.641186 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.641210 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.641222 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.641233 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.641240 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:53Z","lastTransitionTime":"2025-11-23T06:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.743197 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.743233 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.743245 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.743259 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.743269 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:53Z","lastTransitionTime":"2025-11-23T06:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.845191 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.845232 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.845243 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.845257 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.845267 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:53Z","lastTransitionTime":"2025-11-23T06:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.947188 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.947224 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.947232 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.947246 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:53 crc kubenswrapper[4559]: I1123 06:45:53.947256 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:53Z","lastTransitionTime":"2025-11-23T06:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.049487 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.049592 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.049683 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.049744 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.049799 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:54Z","lastTransitionTime":"2025-11-23T06:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.151330 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.151363 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.151371 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.151384 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.151394 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:54Z","lastTransitionTime":"2025-11-23T06:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.253093 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.253129 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.253137 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.253150 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.253165 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:54Z","lastTransitionTime":"2025-11-23T06:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.272928 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.272947 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:54 crc kubenswrapper[4559]: E1123 06:45:54.273078 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:45:54 crc kubenswrapper[4559]: E1123 06:45:54.273175 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.354728 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.354758 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.354769 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.354783 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.354792 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:54Z","lastTransitionTime":"2025-11-23T06:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.456453 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.456512 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.456524 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.456543 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.456556 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:54Z","lastTransitionTime":"2025-11-23T06:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.558863 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.558891 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.558899 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.558909 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.558919 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:54Z","lastTransitionTime":"2025-11-23T06:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.660698 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.660735 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.660749 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.660764 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.660778 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:54Z","lastTransitionTime":"2025-11-23T06:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.762720 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.762758 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.762769 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.762784 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.762794 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:54Z","lastTransitionTime":"2025-11-23T06:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.864499 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.864517 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.864524 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.864535 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.864543 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:54Z","lastTransitionTime":"2025-11-23T06:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.966623 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.966708 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.966724 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.966741 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:54 crc kubenswrapper[4559]: I1123 06:45:54.966755 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:54Z","lastTransitionTime":"2025-11-23T06:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.068372 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.068398 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.068406 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.068416 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.068424 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:55Z","lastTransitionTime":"2025-11-23T06:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.170532 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.170563 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.170572 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.170588 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.170597 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:55Z","lastTransitionTime":"2025-11-23T06:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.272148 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.272171 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.272179 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.272188 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.272195 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:55Z","lastTransitionTime":"2025-11-23T06:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.272528 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.272571 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:45:55 crc kubenswrapper[4559]: E1123 06:45:55.272603 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:45:55 crc kubenswrapper[4559]: E1123 06:45:55.272667 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.373835 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.373860 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.373891 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.373903 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.373914 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:55Z","lastTransitionTime":"2025-11-23T06:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.474916 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.474948 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.474956 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.474968 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.474976 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:55Z","lastTransitionTime":"2025-11-23T06:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.576828 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.576852 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.576860 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.576871 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.576880 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:55Z","lastTransitionTime":"2025-11-23T06:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.679143 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.679176 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.679203 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.679215 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.679224 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:55Z","lastTransitionTime":"2025-11-23T06:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.781254 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.781309 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.781321 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.781335 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.781345 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:55Z","lastTransitionTime":"2025-11-23T06:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.883528 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.883554 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.883562 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.883571 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.883579 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:55Z","lastTransitionTime":"2025-11-23T06:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.985546 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.985577 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.985589 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.985603 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:55 crc kubenswrapper[4559]: I1123 06:45:55.985613 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:55Z","lastTransitionTime":"2025-11-23T06:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.087663 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.087697 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.087705 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.087718 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.087726 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:56Z","lastTransitionTime":"2025-11-23T06:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.189494 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.189524 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.189534 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.189544 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.189553 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:56Z","lastTransitionTime":"2025-11-23T06:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.273546 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:56 crc kubenswrapper[4559]: E1123 06:45:56.273696 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.273807 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:56 crc kubenswrapper[4559]: E1123 06:45:56.273913 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.290632 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.290685 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.290694 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.290705 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.290714 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:56Z","lastTransitionTime":"2025-11-23T06:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.392668 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.392704 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.392714 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.392725 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.392733 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:56Z","lastTransitionTime":"2025-11-23T06:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.495104 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.495129 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.495137 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.495152 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.495159 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:56Z","lastTransitionTime":"2025-11-23T06:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.596371 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.596404 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.596412 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.596422 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.596430 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:56Z","lastTransitionTime":"2025-11-23T06:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.698233 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.698264 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.698277 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.698300 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.698309 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:56Z","lastTransitionTime":"2025-11-23T06:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.800055 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.800087 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.800096 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.800107 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.800115 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:56Z","lastTransitionTime":"2025-11-23T06:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.902290 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.902316 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.902325 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.902335 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:56 crc kubenswrapper[4559]: I1123 06:45:56.902343 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:56Z","lastTransitionTime":"2025-11-23T06:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.003895 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.003922 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.003933 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.003942 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.003950 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:57Z","lastTransitionTime":"2025-11-23T06:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.105582 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.105610 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.105619 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.105631 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.105657 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:57Z","lastTransitionTime":"2025-11-23T06:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.207412 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.207444 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.207452 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.207463 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.207472 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:57Z","lastTransitionTime":"2025-11-23T06:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.273079 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.273130 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:57 crc kubenswrapper[4559]: E1123 06:45:57.273235 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:45:57 crc kubenswrapper[4559]: E1123 06:45:57.273382 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.310213 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.310243 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.310251 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.310267 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.310286 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:57Z","lastTransitionTime":"2025-11-23T06:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.412316 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.412351 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.412362 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.412375 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.412386 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:57Z","lastTransitionTime":"2025-11-23T06:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.514503 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.514541 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.514550 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.514560 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.514570 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:57Z","lastTransitionTime":"2025-11-23T06:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.616636 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.616678 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.616686 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.616696 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.616704 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:57Z","lastTransitionTime":"2025-11-23T06:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.719336 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.719371 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.719382 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.719393 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.719402 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:57Z","lastTransitionTime":"2025-11-23T06:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.821529 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.821575 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.821585 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.821600 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.821609 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:57Z","lastTransitionTime":"2025-11-23T06:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.924165 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.924198 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.924207 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.924224 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:57 crc kubenswrapper[4559]: I1123 06:45:57.924235 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:57Z","lastTransitionTime":"2025-11-23T06:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.026167 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.026198 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.026206 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.026217 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.026226 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:58Z","lastTransitionTime":"2025-11-23T06:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.127917 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.127950 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.127959 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.127991 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.128000 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:58Z","lastTransitionTime":"2025-11-23T06:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.229484 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.229527 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.229542 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.229557 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.229567 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:58Z","lastTransitionTime":"2025-11-23T06:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.272547 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:45:58 crc kubenswrapper[4559]: E1123 06:45:58.272702 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.272851 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:45:58 crc kubenswrapper[4559]: E1123 06:45:58.272895 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.282188 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.290941 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.302438 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0a58330faa93c0acb412514c8429ae62f58ef1a847a66a9a3392736f5d8d88d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0a58330faa93c0acb412514c8429ae62f58ef1a847a66a9a3392736f5d8d88d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:45:43Z\\\",\\\"message\\\":\\\"27.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:43Z is after 2025-08-24T17:21:41Z]\\\\nI1123 06:45:43.858842 6266 services_controller.go:434] Service openshift-operator-lifecycle-manager/catalog-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{catalog-operator-metrics openshift-operator-lifecycle-manager a8e45f62-9ba9-441e-8ed9-b3756498d7e7 4615 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:catalog-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:catalog-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0076825e7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: catalog-operator,},ClusterIP:10.217\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jt56f_openshift-ovn-kubernetes(b12e257e-7887-4795-9221-4db8fd5856dd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.311625 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.320953 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.330872 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.330939 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.330979 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.330997 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.331010 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:58Z","lastTransitionTime":"2025-11-23T06:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.331544 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5bf62d2-3296-4245-970b-2553e647234c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5eb007f4dc7b9bbbed98a8780b4a9ceed7fca2ccd970acd5f7f8a734ae61c55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbebf00280a1ffac1595cd96655356e8567fe98e96ab346e0225f2fcc1c4de9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9k8xh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.340738 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af8237a-8a1f-4d54-aab5-63256e77445d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cceb63202f93e3545d380835e59368e9c8dfb8f818cf6753fd2790574885e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51539e1fed5ea7e5899c119f4a93fd819aaba7ce335225a20b766a4dcbcfcad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://411e9c23598439e930509a324a28db67bc31bad2613822349d7a2fed97937e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a415a92a899290d1b121b3ba83e3d3ca0a84a001256e242cac011a9933fc1cd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a415a92a899290d1b121b3ba83e3d3ca0a84a001256e242cac011a9933fc1cd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.350589 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.358581 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.365433 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.373766 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.383224 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56e33b4cd5df561efe04dcbbf6688c9edca01209b83ca44b40fc61cf562b2e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.389427 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svb6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abb6827e-709d-4ea9-8090-e4f541b1ff3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svb6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.400295 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.408466 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.416080 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.423137 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:58Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.433390 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.433421 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.433429 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.433444 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.433453 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:58Z","lastTransitionTime":"2025-11-23T06:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.535615 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.535687 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.535698 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.535715 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.535727 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:58Z","lastTransitionTime":"2025-11-23T06:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.638914 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.638957 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.638967 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.638983 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.638995 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:58Z","lastTransitionTime":"2025-11-23T06:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.741392 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.741433 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.741441 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.741455 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.741464 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:58Z","lastTransitionTime":"2025-11-23T06:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.843550 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.843584 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.843593 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.843604 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.843613 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:58Z","lastTransitionTime":"2025-11-23T06:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.946350 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.946389 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.946398 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.946413 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:58 crc kubenswrapper[4559]: I1123 06:45:58.946421 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:58Z","lastTransitionTime":"2025-11-23T06:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.048487 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.048525 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.048533 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.048548 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.048558 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:59Z","lastTransitionTime":"2025-11-23T06:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.150736 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.150764 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.150775 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.150786 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.150793 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:59Z","lastTransitionTime":"2025-11-23T06:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.253038 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.253080 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.253089 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.253104 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.253115 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:59Z","lastTransitionTime":"2025-11-23T06:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.273160 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:45:59 crc kubenswrapper[4559]: E1123 06:45:59.273270 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.273387 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:45:59 crc kubenswrapper[4559]: E1123 06:45:59.273515 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.355034 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.355071 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.355080 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.355098 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.355108 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:59Z","lastTransitionTime":"2025-11-23T06:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.456813 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.456841 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.456849 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.456859 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.456867 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:59Z","lastTransitionTime":"2025-11-23T06:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.558722 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.558744 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.558755 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.558769 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.558779 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:59Z","lastTransitionTime":"2025-11-23T06:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.661211 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.661241 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.661260 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.661270 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.661294 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:59Z","lastTransitionTime":"2025-11-23T06:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.763381 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.763431 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.763442 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.763463 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.763478 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:59Z","lastTransitionTime":"2025-11-23T06:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.830408 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.830450 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.830474 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.830492 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.830501 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:59Z","lastTransitionTime":"2025-11-23T06:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:59 crc kubenswrapper[4559]: E1123 06:45:59.842046 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:59Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.844966 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.845017 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.845028 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.845042 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.845050 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:59Z","lastTransitionTime":"2025-11-23T06:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:59 crc kubenswrapper[4559]: E1123 06:45:59.854511 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:59Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.856718 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.856762 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.856772 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.856781 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.856788 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:59Z","lastTransitionTime":"2025-11-23T06:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:59 crc kubenswrapper[4559]: E1123 06:45:59.865755 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:59Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.868159 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.868182 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.868220 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.868233 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.868240 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:59Z","lastTransitionTime":"2025-11-23T06:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:59 crc kubenswrapper[4559]: E1123 06:45:59.876959 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:59Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.879341 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.879435 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.879489 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.879547 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.879599 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:59Z","lastTransitionTime":"2025-11-23T06:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:59 crc kubenswrapper[4559]: E1123 06:45:59.888459 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:59Z is after 2025-08-24T17:21:41Z" Nov 23 06:45:59 crc kubenswrapper[4559]: E1123 06:45:59.888701 4559 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.890320 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.890430 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.890502 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.890576 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.890628 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:59Z","lastTransitionTime":"2025-11-23T06:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.992779 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.992802 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.992810 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.992824 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:45:59 crc kubenswrapper[4559]: I1123 06:45:59.992833 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:45:59Z","lastTransitionTime":"2025-11-23T06:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.094874 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.094897 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.094906 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.094923 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.094930 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:00Z","lastTransitionTime":"2025-11-23T06:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.196434 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.196468 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.196477 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.196487 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.196494 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:00Z","lastTransitionTime":"2025-11-23T06:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.274379 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:00 crc kubenswrapper[4559]: E1123 06:46:00.274466 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.274512 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:00 crc kubenswrapper[4559]: E1123 06:46:00.274546 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.274998 4559 scope.go:117] "RemoveContainer" containerID="f0a58330faa93c0acb412514c8429ae62f58ef1a847a66a9a3392736f5d8d88d" Nov 23 06:46:00 crc kubenswrapper[4559]: E1123 06:46:00.275107 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jt56f_openshift-ovn-kubernetes(b12e257e-7887-4795-9221-4db8fd5856dd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.297753 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.297785 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.297793 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.297802 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.297813 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:00Z","lastTransitionTime":"2025-11-23T06:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.399299 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.399324 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.399333 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.399347 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.399359 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:00Z","lastTransitionTime":"2025-11-23T06:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.501273 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.501304 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.501314 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.501325 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.501333 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:00Z","lastTransitionTime":"2025-11-23T06:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.603164 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.603297 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.603370 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.603434 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.603493 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:00Z","lastTransitionTime":"2025-11-23T06:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.705504 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.705633 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.705828 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.705979 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.706123 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:00Z","lastTransitionTime":"2025-11-23T06:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.807896 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.807950 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.807960 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.807981 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.807999 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:00Z","lastTransitionTime":"2025-11-23T06:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.909750 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.909788 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.909799 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.909812 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:00 crc kubenswrapper[4559]: I1123 06:46:00.909823 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:00Z","lastTransitionTime":"2025-11-23T06:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.011853 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.011885 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.011894 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.011905 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.011914 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:01Z","lastTransitionTime":"2025-11-23T06:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.114274 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.114316 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.114328 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.114343 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.114357 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:01Z","lastTransitionTime":"2025-11-23T06:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.216094 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.216302 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.216362 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.216431 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.216489 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:01Z","lastTransitionTime":"2025-11-23T06:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.272677 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:01 crc kubenswrapper[4559]: E1123 06:46:01.272784 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.272678 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:01 crc kubenswrapper[4559]: E1123 06:46:01.272982 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.318616 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.318656 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.318666 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.318679 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.318690 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:01Z","lastTransitionTime":"2025-11-23T06:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.421050 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.421088 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.421097 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.421110 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.421120 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:01Z","lastTransitionTime":"2025-11-23T06:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.523761 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.523936 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.523994 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.524061 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.524134 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:01Z","lastTransitionTime":"2025-11-23T06:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.625157 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.625287 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.625348 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.625410 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.625481 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:01Z","lastTransitionTime":"2025-11-23T06:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.727214 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.727262 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.727270 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.727283 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.727292 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:01Z","lastTransitionTime":"2025-11-23T06:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.829088 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.829135 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.829145 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.829161 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.829171 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:01Z","lastTransitionTime":"2025-11-23T06:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.930473 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.930517 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.930548 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.930566 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:01 crc kubenswrapper[4559]: I1123 06:46:01.930579 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:01Z","lastTransitionTime":"2025-11-23T06:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.032758 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.032799 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.032808 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.032822 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.032834 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:02Z","lastTransitionTime":"2025-11-23T06:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.134154 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.134264 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.134326 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.134389 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.134452 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:02Z","lastTransitionTime":"2025-11-23T06:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.236149 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.236206 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.236229 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.236243 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.236252 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:02Z","lastTransitionTime":"2025-11-23T06:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.273541 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.273708 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:02 crc kubenswrapper[4559]: E1123 06:46:02.273794 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:02 crc kubenswrapper[4559]: E1123 06:46:02.273921 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.337817 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.337849 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.337857 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.337868 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.337878 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:02Z","lastTransitionTime":"2025-11-23T06:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.439679 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.439700 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.439709 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.439718 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.439725 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:02Z","lastTransitionTime":"2025-11-23T06:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.542409 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.542459 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.542471 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.542489 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.542504 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:02Z","lastTransitionTime":"2025-11-23T06:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.644117 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.644155 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.644164 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.644177 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.644186 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:02Z","lastTransitionTime":"2025-11-23T06:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.746167 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.746197 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.746207 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.746230 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.746239 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:02Z","lastTransitionTime":"2025-11-23T06:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.847841 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.847891 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.847900 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.847911 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.847921 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:02Z","lastTransitionTime":"2025-11-23T06:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.949728 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.949754 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.949763 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.949777 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:02 crc kubenswrapper[4559]: I1123 06:46:02.949788 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:02Z","lastTransitionTime":"2025-11-23T06:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.051743 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.051777 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.051789 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.051803 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.051812 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:03Z","lastTransitionTime":"2025-11-23T06:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.153638 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.153699 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.153708 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.153722 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.153734 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:03Z","lastTransitionTime":"2025-11-23T06:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.256097 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.256148 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.256158 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.256173 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.256185 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:03Z","lastTransitionTime":"2025-11-23T06:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.273516 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.273532 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:03 crc kubenswrapper[4559]: E1123 06:46:03.273609 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:03 crc kubenswrapper[4559]: E1123 06:46:03.273806 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.357532 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.357561 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.357570 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.357582 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.357593 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:03Z","lastTransitionTime":"2025-11-23T06:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.459234 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.459264 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.459273 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.459285 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.459294 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:03Z","lastTransitionTime":"2025-11-23T06:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.561306 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.561334 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.561343 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.561353 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.561362 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:03Z","lastTransitionTime":"2025-11-23T06:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.662846 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.662896 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.662905 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.662917 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.662933 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:03Z","lastTransitionTime":"2025-11-23T06:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.764486 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.764519 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.764527 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.764538 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.764548 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:03Z","lastTransitionTime":"2025-11-23T06:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.866384 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.866415 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.866425 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.866437 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.866446 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:03Z","lastTransitionTime":"2025-11-23T06:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.968529 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.968562 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.968570 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.968582 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:03 crc kubenswrapper[4559]: I1123 06:46:03.968590 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:03Z","lastTransitionTime":"2025-11-23T06:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.070368 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.070400 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.070409 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.070422 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.070430 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:04Z","lastTransitionTime":"2025-11-23T06:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.171679 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.171706 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.171715 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.171728 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.171758 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:04Z","lastTransitionTime":"2025-11-23T06:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.272926 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.272960 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:04 crc kubenswrapper[4559]: E1123 06:46:04.273042 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:04 crc kubenswrapper[4559]: E1123 06:46:04.273177 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.273980 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.274007 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.274032 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.274044 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.274052 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:04Z","lastTransitionTime":"2025-11-23T06:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.375749 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.375785 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.375794 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.375807 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.375816 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:04Z","lastTransitionTime":"2025-11-23T06:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.477444 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.477473 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.477480 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.477491 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.477500 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:04Z","lastTransitionTime":"2025-11-23T06:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.579232 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.579257 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.579265 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.579274 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.579282 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:04Z","lastTransitionTime":"2025-11-23T06:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.680841 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.680866 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.680874 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.680883 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.680891 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:04Z","lastTransitionTime":"2025-11-23T06:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.782577 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.782764 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.782856 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.782916 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.783011 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:04Z","lastTransitionTime":"2025-11-23T06:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.884532 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.884563 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.884571 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.884583 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.884592 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:04Z","lastTransitionTime":"2025-11-23T06:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.986203 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.986225 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.986233 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.986242 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:04 crc kubenswrapper[4559]: I1123 06:46:04.986250 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:04Z","lastTransitionTime":"2025-11-23T06:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.087603 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.087662 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.087673 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.087684 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.087691 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:05Z","lastTransitionTime":"2025-11-23T06:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.188933 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.188968 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.188977 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.188990 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.188998 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:05Z","lastTransitionTime":"2025-11-23T06:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.272553 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:05 crc kubenswrapper[4559]: E1123 06:46:05.272683 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.272557 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:05 crc kubenswrapper[4559]: E1123 06:46:05.273035 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.289898 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.289925 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.289933 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.289944 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.289952 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:05Z","lastTransitionTime":"2025-11-23T06:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.391151 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.391187 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.391196 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.391205 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.391212 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:05Z","lastTransitionTime":"2025-11-23T06:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.492557 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.492593 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.492602 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.492614 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.492623 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:05Z","lastTransitionTime":"2025-11-23T06:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.576364 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/abb6827e-709d-4ea9-8090-e4f541b1ff3f-metrics-certs\") pod \"network-metrics-daemon-svb6z\" (UID: \"abb6827e-709d-4ea9-8090-e4f541b1ff3f\") " pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:05 crc kubenswrapper[4559]: E1123 06:46:05.576461 4559 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:46:05 crc kubenswrapper[4559]: E1123 06:46:05.576502 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/abb6827e-709d-4ea9-8090-e4f541b1ff3f-metrics-certs podName:abb6827e-709d-4ea9-8090-e4f541b1ff3f nodeName:}" failed. No retries permitted until 2025-11-23 06:46:37.576490143 +0000 UTC m=+99.598475756 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/abb6827e-709d-4ea9-8090-e4f541b1ff3f-metrics-certs") pod "network-metrics-daemon-svb6z" (UID: "abb6827e-709d-4ea9-8090-e4f541b1ff3f") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.594482 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.594520 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.594530 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.594544 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.594554 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:05Z","lastTransitionTime":"2025-11-23T06:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.695870 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.695897 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.695905 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.695914 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.695921 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:05Z","lastTransitionTime":"2025-11-23T06:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.797804 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.797831 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.797839 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.797848 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.797855 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:05Z","lastTransitionTime":"2025-11-23T06:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.899370 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.899399 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.899407 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.899419 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:05 crc kubenswrapper[4559]: I1123 06:46:05.899426 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:05Z","lastTransitionTime":"2025-11-23T06:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.000586 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.000616 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.000627 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.000655 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.000664 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:06Z","lastTransitionTime":"2025-11-23T06:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.101820 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.101851 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.101859 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.101871 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.101880 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:06Z","lastTransitionTime":"2025-11-23T06:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.202725 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.202783 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.202794 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.202804 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.202811 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:06Z","lastTransitionTime":"2025-11-23T06:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.273421 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:06 crc kubenswrapper[4559]: E1123 06:46:06.273518 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.273664 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:06 crc kubenswrapper[4559]: E1123 06:46:06.273738 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.304219 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.304286 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.304303 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.304317 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.304325 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:06Z","lastTransitionTime":"2025-11-23T06:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.405881 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.405904 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.405912 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.405922 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.405931 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:06Z","lastTransitionTime":"2025-11-23T06:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.507334 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.507357 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.507365 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.507375 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.507399 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:06Z","lastTransitionTime":"2025-11-23T06:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.608967 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.609006 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.609015 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.609024 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.609031 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:06Z","lastTransitionTime":"2025-11-23T06:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.710423 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.710451 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.710458 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.710468 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.710475 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:06Z","lastTransitionTime":"2025-11-23T06:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.812125 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.812155 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.812172 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.812184 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.812193 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:06Z","lastTransitionTime":"2025-11-23T06:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.913663 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.913695 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.913703 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.913714 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:06 crc kubenswrapper[4559]: I1123 06:46:06.913721 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:06Z","lastTransitionTime":"2025-11-23T06:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.015208 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.015238 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.015252 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.015263 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.015272 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:07Z","lastTransitionTime":"2025-11-23T06:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.116519 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.116543 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.116553 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.116564 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.116571 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:07Z","lastTransitionTime":"2025-11-23T06:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.218286 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.218321 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.218333 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.218346 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.218354 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:07Z","lastTransitionTime":"2025-11-23T06:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.272933 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:07 crc kubenswrapper[4559]: E1123 06:46:07.273008 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.273126 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:07 crc kubenswrapper[4559]: E1123 06:46:07.273232 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.319652 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.319689 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.319698 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.319709 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.319715 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:07Z","lastTransitionTime":"2025-11-23T06:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.421688 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.421715 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.421723 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.421733 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.421742 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:07Z","lastTransitionTime":"2025-11-23T06:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.507272 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qn4h8_18091d80-4837-4894-8583-9a2bd30e10b9/kube-multus/0.log" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.507309 4559 generic.go:334] "Generic (PLEG): container finished" podID="18091d80-4837-4894-8583-9a2bd30e10b9" containerID="aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64" exitCode=1 Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.507330 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qn4h8" event={"ID":"18091d80-4837-4894-8583-9a2bd30e10b9","Type":"ContainerDied","Data":"aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64"} Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.507585 4559 scope.go:117] "RemoveContainer" containerID="aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.516618 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.523392 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.523554 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.523563 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.523576 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.523584 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:07Z","lastTransitionTime":"2025-11-23T06:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.525326 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:46:06Z\\\",\\\"message\\\":\\\"2025-11-23T06:45:21+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9cf01868-bc3f-4119-96d7-74b7770ab8af\\\\n2025-11-23T06:45:21+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9cf01868-bc3f-4119-96d7-74b7770ab8af to /host/opt/cni/bin/\\\\n2025-11-23T06:45:21Z [verbose] multus-daemon started\\\\n2025-11-23T06:45:21Z [verbose] Readiness Indicator file check\\\\n2025-11-23T06:46:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.534616 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56e33b4cd5df561efe04dcbbf6688c9edca01209b83ca44b40fc61cf562b2e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.541458 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svb6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abb6827e-709d-4ea9-8090-e4f541b1ff3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svb6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.551617 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.559449 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.567571 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.576077 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.583811 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.596990 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0a58330faa93c0acb412514c8429ae62f58ef1a847a66a9a3392736f5d8d88d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0a58330faa93c0acb412514c8429ae62f58ef1a847a66a9a3392736f5d8d88d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:45:43Z\\\",\\\"message\\\":\\\"27.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:43Z is after 2025-08-24T17:21:41Z]\\\\nI1123 06:45:43.858842 6266 services_controller.go:434] Service openshift-operator-lifecycle-manager/catalog-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{catalog-operator-metrics openshift-operator-lifecycle-manager a8e45f62-9ba9-441e-8ed9-b3756498d7e7 4615 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:catalog-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:catalog-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0076825e7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: catalog-operator,},ClusterIP:10.217\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jt56f_openshift-ovn-kubernetes(b12e257e-7887-4795-9221-4db8fd5856dd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.606372 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.613131 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.620142 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5bf62d2-3296-4245-970b-2553e647234c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5eb007f4dc7b9bbbed98a8780b4a9ceed7fca2ccd970acd5f7f8a734ae61c55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbebf00280a1ffac1595cd96655356e8567fe98e96ab346e0225f2fcc1c4de9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9k8xh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.625425 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.625458 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.625471 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.625483 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.625491 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:07Z","lastTransitionTime":"2025-11-23T06:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.627453 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af8237a-8a1f-4d54-aab5-63256e77445d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cceb63202f93e3545d380835e59368e9c8dfb8f818cf6753fd2790574885e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51539e1fed5ea7e5899c119f4a93fd819aaba7ce335225a20b766a4dcbcfcad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://411e9c23598439e930509a324a28db67bc31bad2613822349d7a2fed97937e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a415a92a899290d1b121b3ba83e3d3ca0a84a001256e242cac011a9933fc1cd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a415a92a899290d1b121b3ba83e3d3ca0a84a001256e242cac011a9933fc1cd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.636014 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.643252 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.649526 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:07Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.727383 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.727415 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.727426 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.727439 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.727447 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:07Z","lastTransitionTime":"2025-11-23T06:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.829381 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.829414 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.829423 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.829437 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.829445 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:07Z","lastTransitionTime":"2025-11-23T06:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.930496 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.930522 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.930530 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.930539 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:07 crc kubenswrapper[4559]: I1123 06:46:07.930549 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:07Z","lastTransitionTime":"2025-11-23T06:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.032427 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.032457 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.032465 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.032476 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.032483 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:08Z","lastTransitionTime":"2025-11-23T06:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.134117 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.134164 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.134173 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.134185 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.134194 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:08Z","lastTransitionTime":"2025-11-23T06:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.235189 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.235218 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.235226 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.235237 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.235245 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:08Z","lastTransitionTime":"2025-11-23T06:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.273404 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.273438 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:08 crc kubenswrapper[4559]: E1123 06:46:08.273484 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:08 crc kubenswrapper[4559]: E1123 06:46:08.273554 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.284972 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.293325 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.301809 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.309494 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.317877 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:46:06Z\\\",\\\"message\\\":\\\"2025-11-23T06:45:21+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9cf01868-bc3f-4119-96d7-74b7770ab8af\\\\n2025-11-23T06:45:21+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9cf01868-bc3f-4119-96d7-74b7770ab8af to /host/opt/cni/bin/\\\\n2025-11-23T06:45:21Z [verbose] multus-daemon started\\\\n2025-11-23T06:45:21Z [verbose] Readiness Indicator file check\\\\n2025-11-23T06:46:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.326806 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56e33b4cd5df561efe04dcbbf6688c9edca01209b83ca44b40fc61cf562b2e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.333989 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svb6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abb6827e-709d-4ea9-8090-e4f541b1ff3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svb6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.336090 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.336112 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.336121 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.336131 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.336138 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:08Z","lastTransitionTime":"2025-11-23T06:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.341981 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.348953 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.364854 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0a58330faa93c0acb412514c8429ae62f58ef1a847a66a9a3392736f5d8d88d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0a58330faa93c0acb412514c8429ae62f58ef1a847a66a9a3392736f5d8d88d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:45:43Z\\\",\\\"message\\\":\\\"27.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:43Z is after 2025-08-24T17:21:41Z]\\\\nI1123 06:45:43.858842 6266 services_controller.go:434] Service openshift-operator-lifecycle-manager/catalog-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{catalog-operator-metrics openshift-operator-lifecycle-manager a8e45f62-9ba9-441e-8ed9-b3756498d7e7 4615 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:catalog-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:catalog-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0076825e7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: catalog-operator,},ClusterIP:10.217\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jt56f_openshift-ovn-kubernetes(b12e257e-7887-4795-9221-4db8fd5856dd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.372634 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.379131 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.386109 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5bf62d2-3296-4245-970b-2553e647234c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5eb007f4dc7b9bbbed98a8780b4a9ceed7fca2ccd970acd5f7f8a734ae61c55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbebf00280a1ffac1595cd96655356e8567fe98e96ab346e0225f2fcc1c4de9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9k8xh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.392971 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af8237a-8a1f-4d54-aab5-63256e77445d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cceb63202f93e3545d380835e59368e9c8dfb8f818cf6753fd2790574885e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51539e1fed5ea7e5899c119f4a93fd819aaba7ce335225a20b766a4dcbcfcad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://411e9c23598439e930509a324a28db67bc31bad2613822349d7a2fed97937e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a415a92a899290d1b121b3ba83e3d3ca0a84a001256e242cac011a9933fc1cd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a415a92a899290d1b121b3ba83e3d3ca0a84a001256e242cac011a9933fc1cd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.400710 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.408161 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.414357 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.437662 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.437693 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.437702 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.437714 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.437723 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:08Z","lastTransitionTime":"2025-11-23T06:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.510132 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qn4h8_18091d80-4837-4894-8583-9a2bd30e10b9/kube-multus/0.log" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.510195 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qn4h8" event={"ID":"18091d80-4837-4894-8583-9a2bd30e10b9","Type":"ContainerStarted","Data":"73ab4ccc28f84b0b246d2a4464c101f4f3113b2e910eea31f0effd6b89a0a71f"} Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.519669 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.528085 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.535018 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.539029 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.539056 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.539066 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.539078 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.539086 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:08Z","lastTransitionTime":"2025-11-23T06:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.542288 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af8237a-8a1f-4d54-aab5-63256e77445d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cceb63202f93e3545d380835e59368e9c8dfb8f818cf6753fd2790574885e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51539e1fed5ea7e5899c119f4a93fd819aaba7ce335225a20b766a4dcbcfcad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://411e9c23598439e930509a324a28db67bc31bad2613822349d7a2fed97937e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a415a92a899290d1b121b3ba83e3d3ca0a84a001256e242cac011a9933fc1cd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a415a92a899290d1b121b3ba83e3d3ca0a84a001256e242cac011a9933fc1cd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.550514 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.558010 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.565790 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.572772 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.580367 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73ab4ccc28f84b0b246d2a4464c101f4f3113b2e910eea31f0effd6b89a0a71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:46:06Z\\\",\\\"message\\\":\\\"2025-11-23T06:45:21+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9cf01868-bc3f-4119-96d7-74b7770ab8af\\\\n2025-11-23T06:45:21+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9cf01868-bc3f-4119-96d7-74b7770ab8af to /host/opt/cni/bin/\\\\n2025-11-23T06:45:21Z [verbose] multus-daemon started\\\\n2025-11-23T06:45:21Z [verbose] Readiness Indicator file check\\\\n2025-11-23T06:46:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:46:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.589062 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56e33b4cd5df561efe04dcbbf6688c9edca01209b83ca44b40fc61cf562b2e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.595617 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svb6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abb6827e-709d-4ea9-8090-e4f541b1ff3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svb6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.603124 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.610110 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.621713 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0a58330faa93c0acb412514c8429ae62f58ef1a847a66a9a3392736f5d8d88d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0a58330faa93c0acb412514c8429ae62f58ef1a847a66a9a3392736f5d8d88d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:45:43Z\\\",\\\"message\\\":\\\"27.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:43Z is after 2025-08-24T17:21:41Z]\\\\nI1123 06:45:43.858842 6266 services_controller.go:434] Service openshift-operator-lifecycle-manager/catalog-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{catalog-operator-metrics openshift-operator-lifecycle-manager a8e45f62-9ba9-441e-8ed9-b3756498d7e7 4615 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:catalog-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:catalog-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0076825e7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: catalog-operator,},ClusterIP:10.217\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jt56f_openshift-ovn-kubernetes(b12e257e-7887-4795-9221-4db8fd5856dd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.628177 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.635110 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5bf62d2-3296-4245-970b-2553e647234c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5eb007f4dc7b9bbbed98a8780b4a9ceed7fca2ccd970acd5f7f8a734ae61c55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbebf00280a1ffac1595cd96655356e8567fe98e96ab346e0225f2fcc1c4de9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9k8xh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.640714 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.640740 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.640750 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.640762 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.640771 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:08Z","lastTransitionTime":"2025-11-23T06:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.643068 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:08Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.742475 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.742502 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.742510 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.742521 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.742529 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:08Z","lastTransitionTime":"2025-11-23T06:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.844056 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.844098 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.844107 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.844118 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.844125 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:08Z","lastTransitionTime":"2025-11-23T06:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.945673 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.945705 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.945714 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.945726 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:08 crc kubenswrapper[4559]: I1123 06:46:08.945735 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:08Z","lastTransitionTime":"2025-11-23T06:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.049231 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.049264 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.049274 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.049288 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.049297 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:09Z","lastTransitionTime":"2025-11-23T06:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.150887 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.151376 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.151445 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.151508 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.151568 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:09Z","lastTransitionTime":"2025-11-23T06:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.253716 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.253742 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.253750 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.253760 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.253766 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:09Z","lastTransitionTime":"2025-11-23T06:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.273084 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.273102 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:09 crc kubenswrapper[4559]: E1123 06:46:09.273194 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:09 crc kubenswrapper[4559]: E1123 06:46:09.273272 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.355459 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.355488 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.355497 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.355510 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.355535 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:09Z","lastTransitionTime":"2025-11-23T06:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.456768 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.456899 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.456963 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.457017 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.457070 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:09Z","lastTransitionTime":"2025-11-23T06:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.558254 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.558321 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.558332 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.558346 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.558355 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:09Z","lastTransitionTime":"2025-11-23T06:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.659879 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.659904 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.659912 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.659924 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.659933 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:09Z","lastTransitionTime":"2025-11-23T06:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.761904 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.761932 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.761941 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.761952 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.761960 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:09Z","lastTransitionTime":"2025-11-23T06:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.863569 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.863593 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.863603 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.863614 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.863620 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:09Z","lastTransitionTime":"2025-11-23T06:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.951895 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.951989 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.952045 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.952118 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.952188 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:09Z","lastTransitionTime":"2025-11-23T06:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:09 crc kubenswrapper[4559]: E1123 06:46:09.960204 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:09Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.962804 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.962927 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.963001 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.963077 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.963165 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:09Z","lastTransitionTime":"2025-11-23T06:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:09 crc kubenswrapper[4559]: E1123 06:46:09.971446 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:09Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.973336 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.973352 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.973360 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.973370 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.973407 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:09Z","lastTransitionTime":"2025-11-23T06:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:09 crc kubenswrapper[4559]: E1123 06:46:09.980739 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:09Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.982622 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.982674 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.982683 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.982692 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.982698 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:09Z","lastTransitionTime":"2025-11-23T06:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:09 crc kubenswrapper[4559]: E1123 06:46:09.990107 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:09Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.992098 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.992119 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.992156 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.992165 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:09 crc kubenswrapper[4559]: I1123 06:46:09.992172 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:09Z","lastTransitionTime":"2025-11-23T06:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:09 crc kubenswrapper[4559]: E1123 06:46:09.999587 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:09Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:09 crc kubenswrapper[4559]: E1123 06:46:09.999698 4559 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.000610 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.000633 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.000661 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.000670 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.000677 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:10Z","lastTransitionTime":"2025-11-23T06:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.102380 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.102403 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.102411 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.102420 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.102427 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:10Z","lastTransitionTime":"2025-11-23T06:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.203970 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.203995 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.204003 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.204015 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.204022 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:10Z","lastTransitionTime":"2025-11-23T06:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.272655 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:10 crc kubenswrapper[4559]: E1123 06:46:10.272758 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.272890 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:10 crc kubenswrapper[4559]: E1123 06:46:10.272978 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.305256 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.305283 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.305291 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.305304 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.305311 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:10Z","lastTransitionTime":"2025-11-23T06:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.407035 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.407057 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.407065 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.407076 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.407083 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:10Z","lastTransitionTime":"2025-11-23T06:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.508433 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.508461 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.508469 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.508478 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.508486 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:10Z","lastTransitionTime":"2025-11-23T06:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.610241 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.610296 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.610308 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.610321 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.610331 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:10Z","lastTransitionTime":"2025-11-23T06:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.711749 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.711774 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.711782 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.711791 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.711798 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:10Z","lastTransitionTime":"2025-11-23T06:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.813389 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.813420 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.813430 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.813441 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.813449 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:10Z","lastTransitionTime":"2025-11-23T06:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.914937 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.914965 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.914973 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.914998 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:10 crc kubenswrapper[4559]: I1123 06:46:10.915006 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:10Z","lastTransitionTime":"2025-11-23T06:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.016457 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.016485 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.016493 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.016503 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.016511 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:11Z","lastTransitionTime":"2025-11-23T06:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.118070 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.118101 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.118109 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.118122 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.118131 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:11Z","lastTransitionTime":"2025-11-23T06:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.221945 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.221979 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.221988 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.222003 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.222013 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:11Z","lastTransitionTime":"2025-11-23T06:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.272800 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:11 crc kubenswrapper[4559]: E1123 06:46:11.272901 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.272981 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:11 crc kubenswrapper[4559]: E1123 06:46:11.273174 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.323875 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.323917 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.323928 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.323947 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.323965 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:11Z","lastTransitionTime":"2025-11-23T06:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.426240 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.426286 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.426305 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.426321 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.426331 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:11Z","lastTransitionTime":"2025-11-23T06:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.527982 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.528012 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.528021 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.528033 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.528058 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:11Z","lastTransitionTime":"2025-11-23T06:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.630275 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.630312 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.630323 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.630337 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.630349 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:11Z","lastTransitionTime":"2025-11-23T06:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.732226 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.732255 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.732263 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.732274 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.732284 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:11Z","lastTransitionTime":"2025-11-23T06:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.833870 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.833910 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.833919 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.833932 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.833953 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:11Z","lastTransitionTime":"2025-11-23T06:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.936067 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.936094 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.936102 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.936111 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:11 crc kubenswrapper[4559]: I1123 06:46:11.936119 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:11Z","lastTransitionTime":"2025-11-23T06:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.039178 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.039229 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.039239 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.039254 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.039264 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:12Z","lastTransitionTime":"2025-11-23T06:46:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.142050 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.142291 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.142387 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.142497 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.142610 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:12Z","lastTransitionTime":"2025-11-23T06:46:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.244603 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.244667 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.244679 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.244692 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.244700 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:12Z","lastTransitionTime":"2025-11-23T06:46:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.272821 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:12 crc kubenswrapper[4559]: E1123 06:46:12.272934 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.272830 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:12 crc kubenswrapper[4559]: E1123 06:46:12.273311 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.273586 4559 scope.go:117] "RemoveContainer" containerID="f0a58330faa93c0acb412514c8429ae62f58ef1a847a66a9a3392736f5d8d88d" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.346159 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.346293 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.346363 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.346429 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.346488 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:12Z","lastTransitionTime":"2025-11-23T06:46:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.448771 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.448804 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.448815 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.448827 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.448835 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:12Z","lastTransitionTime":"2025-11-23T06:46:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.523421 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jt56f_b12e257e-7887-4795-9221-4db8fd5856dd/ovnkube-controller/2.log" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.525975 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" event={"ID":"b12e257e-7887-4795-9221-4db8fd5856dd","Type":"ContainerStarted","Data":"28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da"} Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.526368 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.546368 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5bf62d2-3296-4245-970b-2553e647234c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5eb007f4dc7b9bbbed98a8780b4a9ceed7fca2ccd970acd5f7f8a734ae61c55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbebf00280a1ffac1595cd96655356e8567fe98e96ab346e0225f2fcc1c4de9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9k8xh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.550764 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.550802 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.550814 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.550831 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.550842 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:12Z","lastTransitionTime":"2025-11-23T06:46:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.558589 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.574750 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.586682 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.594050 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.602952 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af8237a-8a1f-4d54-aab5-63256e77445d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cceb63202f93e3545d380835e59368e9c8dfb8f818cf6753fd2790574885e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51539e1fed5ea7e5899c119f4a93fd819aaba7ce335225a20b766a4dcbcfcad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://411e9c23598439e930509a324a28db67bc31bad2613822349d7a2fed97937e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a415a92a899290d1b121b3ba83e3d3ca0a84a001256e242cac011a9933fc1cd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a415a92a899290d1b121b3ba83e3d3ca0a84a001256e242cac011a9933fc1cd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.611533 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.619333 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.627921 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.636309 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.646045 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73ab4ccc28f84b0b246d2a4464c101f4f3113b2e910eea31f0effd6b89a0a71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:46:06Z\\\",\\\"message\\\":\\\"2025-11-23T06:45:21+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9cf01868-bc3f-4119-96d7-74b7770ab8af\\\\n2025-11-23T06:45:21+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9cf01868-bc3f-4119-96d7-74b7770ab8af to /host/opt/cni/bin/\\\\n2025-11-23T06:45:21Z [verbose] multus-daemon started\\\\n2025-11-23T06:45:21Z [verbose] Readiness Indicator file check\\\\n2025-11-23T06:46:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:46:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.652999 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.653047 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.653058 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.653072 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.653084 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:12Z","lastTransitionTime":"2025-11-23T06:46:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.655611 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56e33b4cd5df561efe04dcbbf6688c9edca01209b83ca44b40fc61cf562b2e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.662839 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svb6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abb6827e-709d-4ea9-8090-e4f541b1ff3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svb6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.671698 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.678868 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.690586 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0a58330faa93c0acb412514c8429ae62f58ef1a847a66a9a3392736f5d8d88d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:45:43Z\\\",\\\"message\\\":\\\"27.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:43Z is after 2025-08-24T17:21:41Z]\\\\nI1123 06:45:43.858842 6266 services_controller.go:434] Service openshift-operator-lifecycle-manager/catalog-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{catalog-operator-metrics openshift-operator-lifecycle-manager a8e45f62-9ba9-441e-8ed9-b3756498d7e7 4615 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:catalog-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:catalog-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0076825e7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: catalog-operator,},ClusterIP:10.217\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:46:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.698166 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:12Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.755552 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.755581 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.755589 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.755602 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.755611 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:12Z","lastTransitionTime":"2025-11-23T06:46:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.857712 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.857759 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.857770 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.857784 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.857792 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:12Z","lastTransitionTime":"2025-11-23T06:46:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.960362 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.960401 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.960410 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.960426 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:12 crc kubenswrapper[4559]: I1123 06:46:12.960436 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:12Z","lastTransitionTime":"2025-11-23T06:46:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.061804 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.061843 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.061853 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.061866 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.061876 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:13Z","lastTransitionTime":"2025-11-23T06:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.163943 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.163972 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.163980 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.163990 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.163999 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:13Z","lastTransitionTime":"2025-11-23T06:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.265720 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.265751 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.265761 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.265772 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.265782 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:13Z","lastTransitionTime":"2025-11-23T06:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.272982 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.273067 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:13 crc kubenswrapper[4559]: E1123 06:46:13.273113 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:13 crc kubenswrapper[4559]: E1123 06:46:13.273233 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.367124 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.367171 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.367183 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.367196 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.367205 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:13Z","lastTransitionTime":"2025-11-23T06:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.469237 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.469272 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.469280 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.469294 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.469336 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:13Z","lastTransitionTime":"2025-11-23T06:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.529774 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jt56f_b12e257e-7887-4795-9221-4db8fd5856dd/ovnkube-controller/3.log" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.530238 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jt56f_b12e257e-7887-4795-9221-4db8fd5856dd/ovnkube-controller/2.log" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.531940 4559 generic.go:334] "Generic (PLEG): container finished" podID="b12e257e-7887-4795-9221-4db8fd5856dd" containerID="28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da" exitCode=1 Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.531971 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" event={"ID":"b12e257e-7887-4795-9221-4db8fd5856dd","Type":"ContainerDied","Data":"28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da"} Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.531999 4559 scope.go:117] "RemoveContainer" containerID="f0a58330faa93c0acb412514c8429ae62f58ef1a847a66a9a3392736f5d8d88d" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.532516 4559 scope.go:117] "RemoveContainer" containerID="28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da" Nov 23 06:46:13 crc kubenswrapper[4559]: E1123 06:46:13.532634 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jt56f_openshift-ovn-kubernetes(b12e257e-7887-4795-9221-4db8fd5856dd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.545687 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73ab4ccc28f84b0b246d2a4464c101f4f3113b2e910eea31f0effd6b89a0a71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:46:06Z\\\",\\\"message\\\":\\\"2025-11-23T06:45:21+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9cf01868-bc3f-4119-96d7-74b7770ab8af\\\\n2025-11-23T06:45:21+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9cf01868-bc3f-4119-96d7-74b7770ab8af to /host/opt/cni/bin/\\\\n2025-11-23T06:45:21Z [verbose] multus-daemon started\\\\n2025-11-23T06:45:21Z [verbose] Readiness Indicator file check\\\\n2025-11-23T06:46:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:46:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.555172 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56e33b4cd5df561efe04dcbbf6688c9edca01209b83ca44b40fc61cf562b2e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.563807 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svb6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abb6827e-709d-4ea9-8090-e4f541b1ff3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svb6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.571185 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.571281 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.571297 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.571332 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.571343 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:13Z","lastTransitionTime":"2025-11-23T06:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.572958 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.581104 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.588664 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.595767 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.603504 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.610390 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.621961 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0a58330faa93c0acb412514c8429ae62f58ef1a847a66a9a3392736f5d8d88d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:45:43Z\\\",\\\"message\\\":\\\"27.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:45:43Z is after 2025-08-24T17:21:41Z]\\\\nI1123 06:45:43.858842 6266 services_controller.go:434] Service openshift-operator-lifecycle-manager/catalog-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{catalog-operator-metrics openshift-operator-lifecycle-manager a8e45f62-9ba9-441e-8ed9-b3756498d7e7 4615 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:catalog-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:catalog-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0076825e7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: catalog-operator,},ClusterIP:10.217\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:46:12Z\\\",\\\"message\\\":\\\"internalTrafficLocal:false, hasNodePort:false}}\\\\nI1123 06:46:12.903821 6670 services_controller.go:445] Built service openshift-dns/dns-default LB template configs for network=default: []services.lbConfig(nil)\\\\nI1123 06:46:12.903206 6670 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc after 0 failed attempt(s)\\\\nI1123 06:46:12.903834 6670 default_network_controller.go:776] Recording success event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1123 06:46:12.903395 6670 services_controller.go:434] Service openshift-apiserver/api retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{api openshift-apiserver 3b54abf8-b632-44a4-b36d-9f489b41a2d2 4787 0 2025-02-23 05:22:52 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[prometheus:openshift-apiserver] map[operator.openshift.io/spec-hash:9c74227d7f96d723d980c50373a5e91f08c5893365bfd5a5040449b1b6585a23 service.alpha.openshift.io/serving-cert-secret-name:serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{apiserver: true,},C\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:46:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.634323 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.641459 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.648828 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5bf62d2-3296-4245-970b-2553e647234c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5eb007f4dc7b9bbbed98a8780b4a9ceed7fca2ccd970acd5f7f8a734ae61c55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbebf00280a1ffac1595cd96655356e8567fe98e96ab346e0225f2fcc1c4de9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9k8xh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.657152 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af8237a-8a1f-4d54-aab5-63256e77445d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cceb63202f93e3545d380835e59368e9c8dfb8f818cf6753fd2790574885e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51539e1fed5ea7e5899c119f4a93fd819aaba7ce335225a20b766a4dcbcfcad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://411e9c23598439e930509a324a28db67bc31bad2613822349d7a2fed97937e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a415a92a899290d1b121b3ba83e3d3ca0a84a001256e242cac011a9933fc1cd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a415a92a899290d1b121b3ba83e3d3ca0a84a001256e242cac011a9933fc1cd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.666147 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.673419 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.673442 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.673450 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.673463 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.673475 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:13Z","lastTransitionTime":"2025-11-23T06:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.675670 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.682720 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:13Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.775854 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.775921 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.775930 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.775945 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.775955 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:13Z","lastTransitionTime":"2025-11-23T06:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.878029 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.878061 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.878070 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.878088 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.878098 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:13Z","lastTransitionTime":"2025-11-23T06:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.980434 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.980486 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.980498 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.980520 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:13 crc kubenswrapper[4559]: I1123 06:46:13.980536 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:13Z","lastTransitionTime":"2025-11-23T06:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.082271 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.082311 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.082321 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.082337 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.082349 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:14Z","lastTransitionTime":"2025-11-23T06:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.184540 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.184662 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.184724 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.184780 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.184839 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:14Z","lastTransitionTime":"2025-11-23T06:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.273491 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.273498 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:14 crc kubenswrapper[4559]: E1123 06:46:14.274270 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:14 crc kubenswrapper[4559]: E1123 06:46:14.274390 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.286332 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.286353 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.286360 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.286372 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.286381 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:14Z","lastTransitionTime":"2025-11-23T06:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.388273 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.388298 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.388307 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.388317 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.388324 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:14Z","lastTransitionTime":"2025-11-23T06:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.489896 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.489918 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.489929 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.489940 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.489948 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:14Z","lastTransitionTime":"2025-11-23T06:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.534508 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jt56f_b12e257e-7887-4795-9221-4db8fd5856dd/ovnkube-controller/3.log" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.536394 4559 scope.go:117] "RemoveContainer" containerID="28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da" Nov 23 06:46:14 crc kubenswrapper[4559]: E1123 06:46:14.536507 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jt56f_openshift-ovn-kubernetes(b12e257e-7887-4795-9221-4db8fd5856dd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.545385 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:14Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.569122 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:14Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.580446 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af8237a-8a1f-4d54-aab5-63256e77445d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cceb63202f93e3545d380835e59368e9c8dfb8f818cf6753fd2790574885e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51539e1fed5ea7e5899c119f4a93fd819aaba7ce335225a20b766a4dcbcfcad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://411e9c23598439e930509a324a28db67bc31bad2613822349d7a2fed97937e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a415a92a899290d1b121b3ba83e3d3ca0a84a001256e242cac011a9933fc1cd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a415a92a899290d1b121b3ba83e3d3ca0a84a001256e242cac011a9933fc1cd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:14Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.589777 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:14Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.591966 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.592010 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.592019 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.592030 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.592040 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:14Z","lastTransitionTime":"2025-11-23T06:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.598665 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:14Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.606758 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:14Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.613839 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:14Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.621595 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73ab4ccc28f84b0b246d2a4464c101f4f3113b2e910eea31f0effd6b89a0a71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:46:06Z\\\",\\\"message\\\":\\\"2025-11-23T06:45:21+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9cf01868-bc3f-4119-96d7-74b7770ab8af\\\\n2025-11-23T06:45:21+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9cf01868-bc3f-4119-96d7-74b7770ab8af to /host/opt/cni/bin/\\\\n2025-11-23T06:45:21Z [verbose] multus-daemon started\\\\n2025-11-23T06:45:21Z [verbose] Readiness Indicator file check\\\\n2025-11-23T06:46:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:46:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:14Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.630299 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56e33b4cd5df561efe04dcbbf6688c9edca01209b83ca44b40fc61cf562b2e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:14Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.636891 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svb6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abb6827e-709d-4ea9-8090-e4f541b1ff3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svb6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:14Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.645250 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:14Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.652084 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:14Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.663560 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:46:12Z\\\",\\\"message\\\":\\\"internalTrafficLocal:false, hasNodePort:false}}\\\\nI1123 06:46:12.903821 6670 services_controller.go:445] Built service openshift-dns/dns-default LB template configs for network=default: []services.lbConfig(nil)\\\\nI1123 06:46:12.903206 6670 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc after 0 failed attempt(s)\\\\nI1123 06:46:12.903834 6670 default_network_controller.go:776] Recording success event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1123 06:46:12.903395 6670 services_controller.go:434] Service openshift-apiserver/api retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{api openshift-apiserver 3b54abf8-b632-44a4-b36d-9f489b41a2d2 4787 0 2025-02-23 05:22:52 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[prometheus:openshift-apiserver] map[operator.openshift.io/spec-hash:9c74227d7f96d723d980c50373a5e91f08c5893365bfd5a5040449b1b6585a23 service.alpha.openshift.io/serving-cert-secret-name:serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{apiserver: true,},C\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:46:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jt56f_openshift-ovn-kubernetes(b12e257e-7887-4795-9221-4db8fd5856dd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:14Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.671157 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:14Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.677935 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5bf62d2-3296-4245-970b-2553e647234c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5eb007f4dc7b9bbbed98a8780b4a9ceed7fca2ccd970acd5f7f8a734ae61c55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbebf00280a1ffac1595cd96655356e8567fe98e96ab346e0225f2fcc1c4de9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9k8xh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:14Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.685356 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:14Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.691480 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:14Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.693445 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.693486 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.693495 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.693505 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.693513 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:14Z","lastTransitionTime":"2025-11-23T06:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.795044 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.795073 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.795082 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.795092 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.795100 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:14Z","lastTransitionTime":"2025-11-23T06:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.896596 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.896637 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.896677 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.896691 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.896700 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:14Z","lastTransitionTime":"2025-11-23T06:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.998831 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.998868 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.998879 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.998894 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:14 crc kubenswrapper[4559]: I1123 06:46:14.998903 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:14Z","lastTransitionTime":"2025-11-23T06:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.100883 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.100915 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.100934 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.100949 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.100958 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:15Z","lastTransitionTime":"2025-11-23T06:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.203168 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.203194 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.203202 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.203211 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.203218 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:15Z","lastTransitionTime":"2025-11-23T06:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.273055 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.273096 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:15 crc kubenswrapper[4559]: E1123 06:46:15.273561 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:15 crc kubenswrapper[4559]: E1123 06:46:15.273685 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.304195 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.304221 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.304230 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.304240 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.304248 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:15Z","lastTransitionTime":"2025-11-23T06:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.405909 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.405935 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.405943 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.405954 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.405963 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:15Z","lastTransitionTime":"2025-11-23T06:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.507455 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.507483 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.507491 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.507502 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.507509 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:15Z","lastTransitionTime":"2025-11-23T06:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.608837 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.608871 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.608881 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.608892 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.608901 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:15Z","lastTransitionTime":"2025-11-23T06:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.711001 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.711042 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.711053 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.711068 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.711078 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:15Z","lastTransitionTime":"2025-11-23T06:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.812837 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.812869 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.812877 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.812889 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.812899 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:15Z","lastTransitionTime":"2025-11-23T06:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.914761 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.914787 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.914795 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.914805 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:15 crc kubenswrapper[4559]: I1123 06:46:15.914813 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:15Z","lastTransitionTime":"2025-11-23T06:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.015903 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.015925 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.015934 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.015944 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.015951 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:16Z","lastTransitionTime":"2025-11-23T06:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.117905 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.117931 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.117941 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.117951 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.117957 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:16Z","lastTransitionTime":"2025-11-23T06:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.219084 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.219111 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.219136 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.219148 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.219157 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:16Z","lastTransitionTime":"2025-11-23T06:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.273033 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.273080 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:16 crc kubenswrapper[4559]: E1123 06:46:16.273145 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:16 crc kubenswrapper[4559]: E1123 06:46:16.273266 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.320813 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.320844 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.320856 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.320867 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.320877 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:16Z","lastTransitionTime":"2025-11-23T06:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.422562 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.422603 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.422613 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.422623 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.422631 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:16Z","lastTransitionTime":"2025-11-23T06:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.524399 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.524429 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.524439 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.524469 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.524479 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:16Z","lastTransitionTime":"2025-11-23T06:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.626719 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.626765 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.626774 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.626787 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.626796 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:16Z","lastTransitionTime":"2025-11-23T06:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.728632 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.728673 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.728681 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.728694 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.728702 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:16Z","lastTransitionTime":"2025-11-23T06:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.830312 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.830337 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.830345 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.830371 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.830379 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:16Z","lastTransitionTime":"2025-11-23T06:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.932197 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.932348 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.932422 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.932491 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:16 crc kubenswrapper[4559]: I1123 06:46:16.932552 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:16Z","lastTransitionTime":"2025-11-23T06:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.034232 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.034270 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.034281 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.034294 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.034303 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:17Z","lastTransitionTime":"2025-11-23T06:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.135455 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.135564 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.135663 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.135740 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.135824 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:17Z","lastTransitionTime":"2025-11-23T06:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.237069 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.237087 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.237094 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.237102 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.237109 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:17Z","lastTransitionTime":"2025-11-23T06:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.272727 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:17 crc kubenswrapper[4559]: E1123 06:46:17.272806 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.272727 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:17 crc kubenswrapper[4559]: E1123 06:46:17.272883 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.338783 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.338813 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.338824 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.338836 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.338844 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:17Z","lastTransitionTime":"2025-11-23T06:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.440507 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.440531 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.440539 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.440549 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.440557 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:17Z","lastTransitionTime":"2025-11-23T06:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.541540 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.541696 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.541775 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.541844 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.541901 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:17Z","lastTransitionTime":"2025-11-23T06:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.643969 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.644017 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.644027 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.644038 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.644045 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:17Z","lastTransitionTime":"2025-11-23T06:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.746406 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.746434 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.746441 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.746453 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.746464 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:17Z","lastTransitionTime":"2025-11-23T06:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.848470 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.848500 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.848508 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.848517 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.848524 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:17Z","lastTransitionTime":"2025-11-23T06:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.950508 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.950535 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.950561 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.950570 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:17 crc kubenswrapper[4559]: I1123 06:46:17.950577 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:17Z","lastTransitionTime":"2025-11-23T06:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.052565 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.052590 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.052598 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.052607 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.052614 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:18Z","lastTransitionTime":"2025-11-23T06:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.154363 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.154389 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.154396 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.154423 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.154431 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:18Z","lastTransitionTime":"2025-11-23T06:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.256095 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.256132 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.256140 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.256150 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.256181 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:18Z","lastTransitionTime":"2025-11-23T06:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.273498 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.273562 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:18 crc kubenswrapper[4559]: E1123 06:46:18.273659 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:18 crc kubenswrapper[4559]: E1123 06:46:18.273822 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.284497 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.292937 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.300567 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.307596 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.315222 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73ab4ccc28f84b0b246d2a4464c101f4f3113b2e910eea31f0effd6b89a0a71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:46:06Z\\\",\\\"message\\\":\\\"2025-11-23T06:45:21+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9cf01868-bc3f-4119-96d7-74b7770ab8af\\\\n2025-11-23T06:45:21+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9cf01868-bc3f-4119-96d7-74b7770ab8af to /host/opt/cni/bin/\\\\n2025-11-23T06:45:21Z [verbose] multus-daemon started\\\\n2025-11-23T06:45:21Z [verbose] Readiness Indicator file check\\\\n2025-11-23T06:46:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:46:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.323881 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56e33b4cd5df561efe04dcbbf6688c9edca01209b83ca44b40fc61cf562b2e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.330578 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svb6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abb6827e-709d-4ea9-8090-e4f541b1ff3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svb6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.339180 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.345783 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.357523 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:46:12Z\\\",\\\"message\\\":\\\"internalTrafficLocal:false, hasNodePort:false}}\\\\nI1123 06:46:12.903821 6670 services_controller.go:445] Built service openshift-dns/dns-default LB template configs for network=default: []services.lbConfig(nil)\\\\nI1123 06:46:12.903206 6670 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc after 0 failed attempt(s)\\\\nI1123 06:46:12.903834 6670 default_network_controller.go:776] Recording success event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1123 06:46:12.903395 6670 services_controller.go:434] Service openshift-apiserver/api retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{api openshift-apiserver 3b54abf8-b632-44a4-b36d-9f489b41a2d2 4787 0 2025-02-23 05:22:52 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[prometheus:openshift-apiserver] map[operator.openshift.io/spec-hash:9c74227d7f96d723d980c50373a5e91f08c5893365bfd5a5040449b1b6585a23 service.alpha.openshift.io/serving-cert-secret-name:serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{apiserver: true,},C\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:46:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jt56f_openshift-ovn-kubernetes(b12e257e-7887-4795-9221-4db8fd5856dd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.358054 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.358090 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.358101 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.358128 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.358139 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:18Z","lastTransitionTime":"2025-11-23T06:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.363906 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.370383 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5bf62d2-3296-4245-970b-2553e647234c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5eb007f4dc7b9bbbed98a8780b4a9ceed7fca2ccd970acd5f7f8a734ae61c55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbebf00280a1ffac1595cd96655356e8567fe98e96ab346e0225f2fcc1c4de9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9k8xh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.378616 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.386519 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.393623 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.399743 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.406595 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af8237a-8a1f-4d54-aab5-63256e77445d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cceb63202f93e3545d380835e59368e9c8dfb8f818cf6753fd2790574885e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51539e1fed5ea7e5899c119f4a93fd819aaba7ce335225a20b766a4dcbcfcad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://411e9c23598439e930509a324a28db67bc31bad2613822349d7a2fed97937e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a415a92a899290d1b121b3ba83e3d3ca0a84a001256e242cac011a9933fc1cd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a415a92a899290d1b121b3ba83e3d3ca0a84a001256e242cac011a9933fc1cd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:18Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.459880 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.459903 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.459911 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.459921 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.459929 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:18Z","lastTransitionTime":"2025-11-23T06:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.563598 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.563673 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.563683 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.563693 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.563700 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:18Z","lastTransitionTime":"2025-11-23T06:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.665042 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.665178 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.665238 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.665302 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.665369 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:18Z","lastTransitionTime":"2025-11-23T06:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.766607 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.766663 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.766674 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.766686 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.766695 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:18Z","lastTransitionTime":"2025-11-23T06:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.867990 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.868016 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.868024 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.868033 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.868041 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:18Z","lastTransitionTime":"2025-11-23T06:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.969416 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.969529 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.969590 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.969673 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:18 crc kubenswrapper[4559]: I1123 06:46:18.969737 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:18Z","lastTransitionTime":"2025-11-23T06:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.070900 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.071017 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.071098 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.071177 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.071241 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:19Z","lastTransitionTime":"2025-11-23T06:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.173161 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.173184 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.173192 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.173200 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.173208 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:19Z","lastTransitionTime":"2025-11-23T06:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.273255 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:19 crc kubenswrapper[4559]: E1123 06:46:19.273356 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.273423 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:19 crc kubenswrapper[4559]: E1123 06:46:19.273539 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.274134 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.274188 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.274198 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.274207 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.274213 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:19Z","lastTransitionTime":"2025-11-23T06:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.376277 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.376304 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.376314 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.376324 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.376331 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:19Z","lastTransitionTime":"2025-11-23T06:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.477839 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.477862 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.477871 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.477880 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.477887 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:19Z","lastTransitionTime":"2025-11-23T06:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.578993 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.579036 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.579046 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.579060 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.579069 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:19Z","lastTransitionTime":"2025-11-23T06:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.680620 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.680660 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.680668 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.680677 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.680685 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:19Z","lastTransitionTime":"2025-11-23T06:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.781777 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.781801 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.781809 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.781817 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.781825 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:19Z","lastTransitionTime":"2025-11-23T06:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.882876 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.882896 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.882904 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.882914 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.882920 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:19Z","lastTransitionTime":"2025-11-23T06:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.984462 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.984569 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.984634 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.984723 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.984781 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:19Z","lastTransitionTime":"2025-11-23T06:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:19 crc kubenswrapper[4559]: I1123 06:46:19.987730 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:19 crc kubenswrapper[4559]: E1123 06:46:19.987848 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:46:19 crc kubenswrapper[4559]: E1123 06:46:19.987866 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:46:19 crc kubenswrapper[4559]: E1123 06:46:19.987875 4559 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:46:19 crc kubenswrapper[4559]: E1123 06:46:19.987910 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-23 06:47:23.987898472 +0000 UTC m=+146.009884086 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.085897 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.085923 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.085931 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.085939 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.085946 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:20Z","lastTransitionTime":"2025-11-23T06:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.088268 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.088349 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:20 crc kubenswrapper[4559]: E1123 06:46:20.088410 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 06:46:20 crc kubenswrapper[4559]: E1123 06:46:20.088421 4559 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 06:46:20 crc kubenswrapper[4559]: E1123 06:46:20.088428 4559 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:46:20 crc kubenswrapper[4559]: E1123 06:46:20.088411 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:24.088399909 +0000 UTC m=+146.110385524 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:46:20 crc kubenswrapper[4559]: E1123 06:46:20.088513 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-23 06:47:24.088505467 +0000 UTC m=+146.110491082 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.188079 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.188119 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.188128 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.188138 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.188147 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:20Z","lastTransitionTime":"2025-11-23T06:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.188679 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.188720 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:20 crc kubenswrapper[4559]: E1123 06:46:20.188781 4559 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:46:20 crc kubenswrapper[4559]: E1123 06:46:20.188826 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:47:24.188815515 +0000 UTC m=+146.210801129 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 06:46:20 crc kubenswrapper[4559]: E1123 06:46:20.188829 4559 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:46:20 crc kubenswrapper[4559]: E1123 06:46:20.188875 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 06:47:24.188861892 +0000 UTC m=+146.210847506 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.205125 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.205184 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.205195 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.205205 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.205212 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:20Z","lastTransitionTime":"2025-11-23T06:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:20 crc kubenswrapper[4559]: E1123 06:46:20.213700 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.216302 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.216384 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.216437 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.216503 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.216556 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:20Z","lastTransitionTime":"2025-11-23T06:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:20 crc kubenswrapper[4559]: E1123 06:46:20.224228 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.226443 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.226533 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.226596 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.226679 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.226733 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:20Z","lastTransitionTime":"2025-11-23T06:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:20 crc kubenswrapper[4559]: E1123 06:46:20.234782 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.237206 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.237296 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.237364 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.237429 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.237479 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:20Z","lastTransitionTime":"2025-11-23T06:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:20 crc kubenswrapper[4559]: E1123 06:46:20.244876 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.246880 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.246906 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.246914 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.246924 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.246931 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:20Z","lastTransitionTime":"2025-11-23T06:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:20 crc kubenswrapper[4559]: E1123 06:46:20.254563 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:20Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:20 crc kubenswrapper[4559]: E1123 06:46:20.254697 4559 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.272832 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.272868 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:20 crc kubenswrapper[4559]: E1123 06:46:20.272936 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:20 crc kubenswrapper[4559]: E1123 06:46:20.273065 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.289266 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.289351 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.289419 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.289476 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.289532 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:20Z","lastTransitionTime":"2025-11-23T06:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.391805 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.391828 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.391837 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.391850 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.391858 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:20Z","lastTransitionTime":"2025-11-23T06:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.493703 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.493739 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.493749 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.493761 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.493782 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:20Z","lastTransitionTime":"2025-11-23T06:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.595512 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.595530 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.595544 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.595553 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.595559 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:20Z","lastTransitionTime":"2025-11-23T06:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.697249 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.697281 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.697290 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.697302 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.697311 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:20Z","lastTransitionTime":"2025-11-23T06:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.799123 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.799175 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.799184 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.799196 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.799204 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:20Z","lastTransitionTime":"2025-11-23T06:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.901194 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.901219 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.901227 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.901239 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:20 crc kubenswrapper[4559]: I1123 06:46:20.901247 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:20Z","lastTransitionTime":"2025-11-23T06:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.003006 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.003041 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.003048 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.003058 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.003067 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:21Z","lastTransitionTime":"2025-11-23T06:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.104814 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.104875 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.104883 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.104895 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.104903 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:21Z","lastTransitionTime":"2025-11-23T06:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.206347 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.206374 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.206384 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.206394 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.206402 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:21Z","lastTransitionTime":"2025-11-23T06:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.273491 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:21 crc kubenswrapper[4559]: E1123 06:46:21.273636 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.273499 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:21 crc kubenswrapper[4559]: E1123 06:46:21.273855 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.308158 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.308210 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.308220 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.308232 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.308241 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:21Z","lastTransitionTime":"2025-11-23T06:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.409335 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.409356 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.409363 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.409372 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.409379 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:21Z","lastTransitionTime":"2025-11-23T06:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.511126 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.511147 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.511154 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.511163 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.511171 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:21Z","lastTransitionTime":"2025-11-23T06:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.612385 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.612411 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.612419 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.612428 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.612435 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:21Z","lastTransitionTime":"2025-11-23T06:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.714306 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.714332 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.714339 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.714347 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.714356 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:21Z","lastTransitionTime":"2025-11-23T06:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.815936 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.815956 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.815964 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.815973 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.815979 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:21Z","lastTransitionTime":"2025-11-23T06:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.917671 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.917699 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.917709 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.917721 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:21 crc kubenswrapper[4559]: I1123 06:46:21.917729 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:21Z","lastTransitionTime":"2025-11-23T06:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.019877 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.019918 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.019929 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.019943 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.019954 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:22Z","lastTransitionTime":"2025-11-23T06:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.121579 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.121610 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.121618 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.121630 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.121638 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:22Z","lastTransitionTime":"2025-11-23T06:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.223758 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.223797 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.223806 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.223818 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.223826 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:22Z","lastTransitionTime":"2025-11-23T06:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.273444 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:22 crc kubenswrapper[4559]: E1123 06:46:22.273545 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.273562 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:22 crc kubenswrapper[4559]: E1123 06:46:22.273668 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.325699 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.325724 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.325733 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.325743 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.325752 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:22Z","lastTransitionTime":"2025-11-23T06:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.427685 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.427708 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.427716 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.427725 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.427734 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:22Z","lastTransitionTime":"2025-11-23T06:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.529722 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.529756 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.529764 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.529776 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.529784 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:22Z","lastTransitionTime":"2025-11-23T06:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.631575 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.631606 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.631616 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.631629 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.631638 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:22Z","lastTransitionTime":"2025-11-23T06:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.733733 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.733769 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.733781 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.733795 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.733803 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:22Z","lastTransitionTime":"2025-11-23T06:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.835029 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.835055 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.835062 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.835072 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.835080 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:22Z","lastTransitionTime":"2025-11-23T06:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.936629 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.936680 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.936692 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.936703 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:22 crc kubenswrapper[4559]: I1123 06:46:22.936711 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:22Z","lastTransitionTime":"2025-11-23T06:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.038264 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.038396 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.038404 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.038413 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.038420 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:23Z","lastTransitionTime":"2025-11-23T06:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.139760 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.139789 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.139798 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.139810 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.139818 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:23Z","lastTransitionTime":"2025-11-23T06:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.241318 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.241354 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.241363 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.241374 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.241382 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:23Z","lastTransitionTime":"2025-11-23T06:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.273541 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.273596 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:23 crc kubenswrapper[4559]: E1123 06:46:23.273626 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:23 crc kubenswrapper[4559]: E1123 06:46:23.273707 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.342617 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.342659 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.342669 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.342680 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.342687 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:23Z","lastTransitionTime":"2025-11-23T06:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.444247 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.444274 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.444300 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.444311 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.444319 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:23Z","lastTransitionTime":"2025-11-23T06:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.545909 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.545946 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.545956 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.545968 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.545979 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:23Z","lastTransitionTime":"2025-11-23T06:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.647552 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.647580 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.647588 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.647597 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.647605 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:23Z","lastTransitionTime":"2025-11-23T06:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.749135 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.749172 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.749179 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.749189 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.749197 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:23Z","lastTransitionTime":"2025-11-23T06:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.851074 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.851113 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.851121 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.851133 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.851141 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:23Z","lastTransitionTime":"2025-11-23T06:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.952733 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.952766 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.952773 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.952785 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:23 crc kubenswrapper[4559]: I1123 06:46:23.952793 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:23Z","lastTransitionTime":"2025-11-23T06:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.054360 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.054384 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.054392 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.054402 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.054410 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:24Z","lastTransitionTime":"2025-11-23T06:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.156218 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.156246 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.156253 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.156264 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.156271 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:24Z","lastTransitionTime":"2025-11-23T06:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.257527 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.257550 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.257557 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.257566 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.257573 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:24Z","lastTransitionTime":"2025-11-23T06:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.273017 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:24 crc kubenswrapper[4559]: E1123 06:46:24.273119 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.273021 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:24 crc kubenswrapper[4559]: E1123 06:46:24.273188 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.359138 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.359165 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.359173 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.359182 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.359188 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:24Z","lastTransitionTime":"2025-11-23T06:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.460982 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.461013 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.461021 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.461032 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.461041 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:24Z","lastTransitionTime":"2025-11-23T06:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.562396 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.562420 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.562428 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.562437 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.562443 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:24Z","lastTransitionTime":"2025-11-23T06:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.664243 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.664270 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.664279 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.664289 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.664298 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:24Z","lastTransitionTime":"2025-11-23T06:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.765543 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.765718 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.765798 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.765873 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.765934 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:24Z","lastTransitionTime":"2025-11-23T06:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.867725 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.867751 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.867759 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.867768 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.867776 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:24Z","lastTransitionTime":"2025-11-23T06:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.969612 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.969638 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.969666 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.969676 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:24 crc kubenswrapper[4559]: I1123 06:46:24.969682 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:24Z","lastTransitionTime":"2025-11-23T06:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.070780 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.070805 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.070814 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.070825 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.070832 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:25Z","lastTransitionTime":"2025-11-23T06:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.172815 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.172844 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.172852 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.172862 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.172871 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:25Z","lastTransitionTime":"2025-11-23T06:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.273219 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.273275 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:25 crc kubenswrapper[4559]: E1123 06:46:25.273354 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:25 crc kubenswrapper[4559]: E1123 06:46:25.273534 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.274060 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.274099 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.274108 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.274118 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.274126 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:25Z","lastTransitionTime":"2025-11-23T06:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.375181 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.375211 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.375221 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.375235 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.375248 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:25Z","lastTransitionTime":"2025-11-23T06:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.476718 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.476748 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.476757 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.476766 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.476773 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:25Z","lastTransitionTime":"2025-11-23T06:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.578227 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.578255 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.578263 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.578274 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.578283 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:25Z","lastTransitionTime":"2025-11-23T06:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.680275 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.680302 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.680311 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.680321 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.680328 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:25Z","lastTransitionTime":"2025-11-23T06:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.781914 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.781943 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.781951 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.781962 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.781969 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:25Z","lastTransitionTime":"2025-11-23T06:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.883706 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.883743 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.883751 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.883760 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.883768 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:25Z","lastTransitionTime":"2025-11-23T06:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.985274 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.985298 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.985306 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.985317 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:25 crc kubenswrapper[4559]: I1123 06:46:25.985328 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:25Z","lastTransitionTime":"2025-11-23T06:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.086571 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.086711 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.086780 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.086856 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.086924 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:26Z","lastTransitionTime":"2025-11-23T06:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.188813 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.188835 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.188842 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.188851 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.188859 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:26Z","lastTransitionTime":"2025-11-23T06:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.272708 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.272830 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:26 crc kubenswrapper[4559]: E1123 06:46:26.273001 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:26 crc kubenswrapper[4559]: E1123 06:46:26.273114 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.290049 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.290087 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.290097 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.290108 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.290118 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:26Z","lastTransitionTime":"2025-11-23T06:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.392093 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.392122 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.392130 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.392140 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.392148 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:26Z","lastTransitionTime":"2025-11-23T06:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.493807 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.493835 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.493843 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.493852 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.493859 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:26Z","lastTransitionTime":"2025-11-23T06:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.595790 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.595818 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.595826 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.595835 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.595844 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:26Z","lastTransitionTime":"2025-11-23T06:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.697483 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.697505 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.697513 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.697522 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.697530 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:26Z","lastTransitionTime":"2025-11-23T06:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.798772 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.798888 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.798962 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.799021 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.799085 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:26Z","lastTransitionTime":"2025-11-23T06:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.900550 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.900573 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.900580 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.900590 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:26 crc kubenswrapper[4559]: I1123 06:46:26.900597 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:26Z","lastTransitionTime":"2025-11-23T06:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.001975 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.002005 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.002012 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.002024 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.002032 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:27Z","lastTransitionTime":"2025-11-23T06:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.103808 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.103834 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.103842 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.103853 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.103862 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:27Z","lastTransitionTime":"2025-11-23T06:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.205460 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.205493 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.205503 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.205514 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.205523 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:27Z","lastTransitionTime":"2025-11-23T06:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.273550 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.273558 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:27 crc kubenswrapper[4559]: E1123 06:46:27.273945 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:27 crc kubenswrapper[4559]: E1123 06:46:27.274044 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.280841 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.307119 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.307213 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.307282 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.307341 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.307405 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:27Z","lastTransitionTime":"2025-11-23T06:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.408845 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.408875 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.408883 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.408895 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.408904 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:27Z","lastTransitionTime":"2025-11-23T06:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.510416 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.510447 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.510457 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.510468 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.510476 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:27Z","lastTransitionTime":"2025-11-23T06:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.612532 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.612555 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.612562 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.612572 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.612578 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:27Z","lastTransitionTime":"2025-11-23T06:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.714299 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.714343 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.714352 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.714361 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.714368 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:27Z","lastTransitionTime":"2025-11-23T06:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.816085 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.816113 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.816122 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.816133 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.816141 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:27Z","lastTransitionTime":"2025-11-23T06:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.917911 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.917941 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.917950 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.917962 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:27 crc kubenswrapper[4559]: I1123 06:46:27.917970 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:27Z","lastTransitionTime":"2025-11-23T06:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.019811 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.019841 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.019848 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.019856 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.019863 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:28Z","lastTransitionTime":"2025-11-23T06:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.121010 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.121143 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.121208 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.121279 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.121345 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:28Z","lastTransitionTime":"2025-11-23T06:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.222546 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.222601 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.222612 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.222627 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.222658 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:28Z","lastTransitionTime":"2025-11-23T06:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.273396 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.273436 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:28 crc kubenswrapper[4559]: E1123 06:46:28.273815 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:28 crc kubenswrapper[4559]: E1123 06:46:28.273877 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.282780 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.290118 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4731beee-0cac-4189-8a70-743b0b709095\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://329b28a0be8f9704f34657380b6d9eb9f6162c09f621ff76328f285b090ce03a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bd57x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-s7zkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.301714 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12e257e-7887-4795-9221-4db8fd5856dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:46:12Z\\\",\\\"message\\\":\\\"internalTrafficLocal:false, hasNodePort:false}}\\\\nI1123 06:46:12.903821 6670 services_controller.go:445] Built service openshift-dns/dns-default LB template configs for network=default: []services.lbConfig(nil)\\\\nI1123 06:46:12.903206 6670 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc after 0 failed attempt(s)\\\\nI1123 06:46:12.903834 6670 default_network_controller.go:776] Recording success event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1123 06:46:12.903395 6670 services_controller.go:434] Service openshift-apiserver/api retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{api openshift-apiserver 3b54abf8-b632-44a4-b36d-9f489b41a2d2 4787 0 2025-02-23 05:22:52 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[prometheus:openshift-apiserver] map[operator.openshift.io/spec-hash:9c74227d7f96d723d980c50373a5e91f08c5893365bfd5a5040449b1b6585a23 service.alpha.openshift.io/serving-cert-secret-name:serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{apiserver: true,},C\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:46:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jt56f_openshift-ovn-kubernetes(b12e257e-7887-4795-9221-4db8fd5856dd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jcdbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jt56f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.309451 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c51f2a51ab4d6b3c7ab54144e9816c14abec8ad5c1488dff0248c33c19ae8691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1777f40d2cf2c381992b5e055a495fbc2ba778a46ef2abc15ae5b1b57d271678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.315403 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njqzh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"16844d77-d909-4a27-bc7b-f99e97662ba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96773f54497511a7744484c08463efe32838f63fbaf2943ee81d1f14b566b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvpfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njqzh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.321784 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5bf62d2-3296-4245-970b-2553e647234c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5eb007f4dc7b9bbbed98a8780b4a9ceed7fca2ccd970acd5f7f8a734ae61c55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbebf00280a1ffac1595cd96655356e8567fe98e96ab346e0225f2fcc1c4de9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wl8jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9k8xh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.323631 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.323681 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.323689 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.323699 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.323707 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:28Z","lastTransitionTime":"2025-11-23T06:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.329044 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5af8237a-8a1f-4d54-aab5-63256e77445d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7cceb63202f93e3545d380835e59368e9c8dfb8f818cf6753fd2790574885e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51539e1fed5ea7e5899c119f4a93fd819aaba7ce335225a20b766a4dcbcfcad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://411e9c23598439e930509a324a28db67bc31bad2613822349d7a2fed97937e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a415a92a899290d1b121b3ba83e3d3ca0a84a001256e242cac011a9933fc1cd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a415a92a899290d1b121b3ba83e3d3ca0a84a001256e242cac011a9933fc1cd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.338628 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87392c8ba0f3161ee7853a244615808bfc31f36dc1fcc0418a3f239a23f4ea0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.346320 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.352272 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w5tlv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdf91a44-8c01-4104-8d6f-e76b8c321166\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e9949d48b3a729bcd3732aa7a36f91c68c8ab8665408ccd69fa2fb0ebabf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7g4kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w5tlv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.361223 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9164873-0a60-498c-9b37-cd73817b066a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56e33b4cd5df561efe04dcbbf6688c9edca01209b83ca44b40fc61cf562b2e82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce020a6e086cac6eed3334df41b610f0aeb0d02cb6386c71eac37dbdd5327c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f726c2b672ac6aee201612c76b06e7263b1c46c0e1210e311dc96fef800bfbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0da7a240bcd89c652d44b57a88b422af670f978961d0b35d57a9867d0123c125\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ebd3548fffd8029ff68b4a3b59aadac3e5ec8479e234fe79a7d6d91e043cd74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7de4fc52e444810184bfc55e14332d0046b1e2e794e133bc3cf75f33472841be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://878bcb8b39938689f340f717c11e36811b5e0c4209a66b0dd043f77cccaa8c36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:45:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n7s7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rlxnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.367708 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-svb6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abb6827e-709d-4ea9-8090-e4f541b1ff3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4xp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-svb6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.374046 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4aa38ed-7117-4840-a277-1531979556bd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://911f55a9468868f66060d4080ebf586aa360927045d4917667d2cb186aac82fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593cf113905c89a79b8db618d0011376eb1db9755e6580e56c16b3af3c358675\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://593cf113905c89a79b8db618d0011376eb1db9755e6580e56c16b3af3c358675\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.382402 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3599e095-5776-4472-85ae-39e05562dc4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad70bf342eb152d8a126097e9123e1e4f266400b0394ec3631e9b31beba642b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbc6945182131967605bcbce718b0a36c7e62f3e9283d8389d76109195d68ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f7be614aa0b1bd479d7cbf616327249eae83d82e58ae66d76d920fb47310b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72549cfa335b148dc47851a3a347f53096009b34412ee47598c972f92ae423ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e11708e70b307ac7d0457f7cd57256e8cf75f1cb8c8bb1e4bcd47b9a181d4a00\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T06:45:15Z\\\",\\\"message\\\":\\\"le observer\\\\nW1123 06:45:15.268682 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1123 06:45:15.268773 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1123 06:45:15.270160 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4218581297/tls.crt::/tmp/serving-cert-4218581297/tls.key\\\\\\\"\\\\nI1123 06:45:15.426554 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1123 06:45:15.428730 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1123 06:45:15.428746 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1123 06:45:15.428765 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1123 06:45:15.428771 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1123 06:45:15.433304 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1123 06:45:15.433324 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1123 06:45:15.433326 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1123 06:45:15.433330 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1123 06:45:15.433351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1123 06:45:15.433354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1123 06:45:15.433357 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1123 06:45:15.433359 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1123 06:45:15.434532 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3528309d78853da4fef585e43a4bed5c5da453422c464c565b30cf506609b85c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4549b62b2814c9b3c3dead41c7a155e9daefb68c04e17d1279c9c5267c8cd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T06:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.389821 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45a2e49b-b89f-4c52-8e2d-3344b1ccb3c7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec70eaaf106b40cdda364bd862b6654bec3f48a61433e2f8cf2d691ff947f9c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5471d02b4c037e03c954084bd0218f68e131f7520ed2239304d3ef4a4ee2010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2893ee3e9854265197f0dedbe3786c6ce4466ae1dd982f792a7756ada8165d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5debae8fac8fae56e180d608aa4f0e6c04639ff5ff6052ffa287722f131e4a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.397131 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.403770 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6596783049a796d88368eaba18223de617046244189696160ec90a4e496090e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.411016 4559 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qn4h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18091d80-4837-4894-8583-9a2bd30e10b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73ab4ccc28f84b0b246d2a4464c101f4f3113b2e910eea31f0effd6b89a0a71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T06:46:06Z\\\",\\\"message\\\":\\\"2025-11-23T06:45:21+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9cf01868-bc3f-4119-96d7-74b7770ab8af\\\\n2025-11-23T06:45:21+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9cf01868-bc3f-4119-96d7-74b7770ab8af to /host/opt/cni/bin/\\\\n2025-11-23T06:45:21Z [verbose] multus-daemon started\\\\n2025-11-23T06:45:21Z [verbose] Readiness Indicator file check\\\\n2025-11-23T06:46:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T06:45:21Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T06:46:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtqzh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T06:45:20Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qn4h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:28Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.425187 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.425212 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.425220 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.425231 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.425239 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:28Z","lastTransitionTime":"2025-11-23T06:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.526909 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.526936 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.526943 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.526953 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.526961 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:28Z","lastTransitionTime":"2025-11-23T06:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.628627 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.628675 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.628687 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.628699 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.628708 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:28Z","lastTransitionTime":"2025-11-23T06:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.729769 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.729800 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.729809 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.729822 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.729834 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:28Z","lastTransitionTime":"2025-11-23T06:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.831084 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.831107 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.831115 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.831124 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.831132 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:28Z","lastTransitionTime":"2025-11-23T06:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.932780 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.932887 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.932945 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.933011 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:28 crc kubenswrapper[4559]: I1123 06:46:28.933080 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:28Z","lastTransitionTime":"2025-11-23T06:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.034670 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.034696 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.034704 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.034714 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.034722 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:29Z","lastTransitionTime":"2025-11-23T06:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.135884 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.135910 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.135920 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.135929 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.135936 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:29Z","lastTransitionTime":"2025-11-23T06:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.237435 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.237542 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.237606 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.237688 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.237748 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:29Z","lastTransitionTime":"2025-11-23T06:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.273250 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.273257 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:29 crc kubenswrapper[4559]: E1123 06:46:29.273359 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:29 crc kubenswrapper[4559]: E1123 06:46:29.273715 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.273917 4559 scope.go:117] "RemoveContainer" containerID="28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da" Nov 23 06:46:29 crc kubenswrapper[4559]: E1123 06:46:29.274034 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jt56f_openshift-ovn-kubernetes(b12e257e-7887-4795-9221-4db8fd5856dd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.338976 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.339028 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.339050 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.339075 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.339083 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:29Z","lastTransitionTime":"2025-11-23T06:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.440637 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.440676 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.440684 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.440696 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.440703 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:29Z","lastTransitionTime":"2025-11-23T06:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.541945 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.541969 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.541976 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.541985 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.541992 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:29Z","lastTransitionTime":"2025-11-23T06:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.643660 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.643686 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.643694 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.643703 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.643710 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:29Z","lastTransitionTime":"2025-11-23T06:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.744973 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.744998 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.745007 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.745016 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.745024 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:29Z","lastTransitionTime":"2025-11-23T06:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.846909 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.846941 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.846951 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.846962 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.846970 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:29Z","lastTransitionTime":"2025-11-23T06:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.948854 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.948885 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.948894 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.948904 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:29 crc kubenswrapper[4559]: I1123 06:46:29.948912 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:29Z","lastTransitionTime":"2025-11-23T06:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.050020 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.050093 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.050106 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.050119 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.050129 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:30Z","lastTransitionTime":"2025-11-23T06:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.151307 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.151337 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.151345 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.151355 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.151362 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:30Z","lastTransitionTime":"2025-11-23T06:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.253229 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.253253 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.253261 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.253272 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.253278 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:30Z","lastTransitionTime":"2025-11-23T06:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.272775 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.272795 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:30 crc kubenswrapper[4559]: E1123 06:46:30.272865 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:30 crc kubenswrapper[4559]: E1123 06:46:30.272935 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.354614 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.354662 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.354673 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.354684 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.354693 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:30Z","lastTransitionTime":"2025-11-23T06:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.455827 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.455855 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.455864 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.455873 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.455881 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:30Z","lastTransitionTime":"2025-11-23T06:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.505070 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.505098 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.505105 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.505113 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.505121 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:30Z","lastTransitionTime":"2025-11-23T06:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:30 crc kubenswrapper[4559]: E1123 06:46:30.513994 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.516200 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.516227 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.516234 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.516245 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.516252 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:30Z","lastTransitionTime":"2025-11-23T06:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:30 crc kubenswrapper[4559]: E1123 06:46:30.523634 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.525702 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.525729 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.525738 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.525749 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.525757 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:30Z","lastTransitionTime":"2025-11-23T06:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:30 crc kubenswrapper[4559]: E1123 06:46:30.532946 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.534770 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.534795 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.534805 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.534815 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.534821 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:30Z","lastTransitionTime":"2025-11-23T06:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:30 crc kubenswrapper[4559]: E1123 06:46:30.542318 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.544075 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.544101 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.544109 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.544119 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.544126 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:30Z","lastTransitionTime":"2025-11-23T06:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:30 crc kubenswrapper[4559]: E1123 06:46:30.551139 4559 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T06:46:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5986746-4ff2-4418-9747-f60dccf1cb6d\\\",\\\"systemUUID\\\":\\\"8b1b3dc3-1aa7-4149-9ed7-13994feca6e6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T06:46:30Z is after 2025-08-24T17:21:41Z" Nov 23 06:46:30 crc kubenswrapper[4559]: E1123 06:46:30.551239 4559 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.556867 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.556891 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.556902 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.556911 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.556934 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:30Z","lastTransitionTime":"2025-11-23T06:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.658310 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.658336 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.658344 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.658354 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.658361 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:30Z","lastTransitionTime":"2025-11-23T06:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.760339 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.760365 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.760375 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.760388 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.760396 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:30Z","lastTransitionTime":"2025-11-23T06:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.862061 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.862087 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.862095 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.862105 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.862113 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:30Z","lastTransitionTime":"2025-11-23T06:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.963125 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.963159 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.963168 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.963177 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:30 crc kubenswrapper[4559]: I1123 06:46:30.963183 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:30Z","lastTransitionTime":"2025-11-23T06:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.064592 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.064617 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.064660 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.064683 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.064692 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:31Z","lastTransitionTime":"2025-11-23T06:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.166116 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.166141 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.166149 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.166159 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.166167 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:31Z","lastTransitionTime":"2025-11-23T06:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.267853 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.267881 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.267891 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.267902 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.267910 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:31Z","lastTransitionTime":"2025-11-23T06:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.273163 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:31 crc kubenswrapper[4559]: E1123 06:46:31.273240 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.273175 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.273299 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:31 crc kubenswrapper[4559]: E1123 06:46:31.273368 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:31 crc kubenswrapper[4559]: E1123 06:46:31.273468 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.369066 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.369093 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.369101 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.369129 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.369137 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:31Z","lastTransitionTime":"2025-11-23T06:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.470476 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.470498 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.470507 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.470516 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.470525 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:31Z","lastTransitionTime":"2025-11-23T06:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.572158 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.572183 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.572192 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.572201 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.572208 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:31Z","lastTransitionTime":"2025-11-23T06:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.673813 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.673836 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.673843 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.673852 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.673860 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:31Z","lastTransitionTime":"2025-11-23T06:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.775551 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.775577 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.775585 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.775598 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.775605 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:31Z","lastTransitionTime":"2025-11-23T06:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.876805 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.876838 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.876846 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.876857 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.876864 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:31Z","lastTransitionTime":"2025-11-23T06:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.978981 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.979011 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.979019 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.979031 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:31 crc kubenswrapper[4559]: I1123 06:46:31.979038 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:31Z","lastTransitionTime":"2025-11-23T06:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.080452 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.080611 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.080704 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.080772 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.080830 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:32Z","lastTransitionTime":"2025-11-23T06:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.182875 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.182914 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.182922 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.182932 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.182940 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:32Z","lastTransitionTime":"2025-11-23T06:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.273485 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:32 crc kubenswrapper[4559]: E1123 06:46:32.273583 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.284899 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.284936 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.284946 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.284958 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.284967 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:32Z","lastTransitionTime":"2025-11-23T06:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.386411 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.386435 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.386444 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.386453 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.386460 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:32Z","lastTransitionTime":"2025-11-23T06:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.488040 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.488078 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.488088 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.488098 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.488106 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:32Z","lastTransitionTime":"2025-11-23T06:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.589991 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.590018 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.590028 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.590036 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.590053 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:32Z","lastTransitionTime":"2025-11-23T06:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.691943 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.691975 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.691985 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.691998 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.692007 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:32Z","lastTransitionTime":"2025-11-23T06:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.794050 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.794075 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.794083 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.794091 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.794098 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:32Z","lastTransitionTime":"2025-11-23T06:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.897021 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.897050 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.897059 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.897069 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.897092 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:32Z","lastTransitionTime":"2025-11-23T06:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.998987 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.999009 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.999017 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.999026 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:32 crc kubenswrapper[4559]: I1123 06:46:32.999035 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:32Z","lastTransitionTime":"2025-11-23T06:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.101279 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.101321 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.101332 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.101343 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.101350 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:33Z","lastTransitionTime":"2025-11-23T06:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.202564 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.202586 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.202594 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.202604 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.202611 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:33Z","lastTransitionTime":"2025-11-23T06:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.272896 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:33 crc kubenswrapper[4559]: E1123 06:46:33.272982 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.273109 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.273114 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:33 crc kubenswrapper[4559]: E1123 06:46:33.273163 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:33 crc kubenswrapper[4559]: E1123 06:46:33.273341 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.304752 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.304778 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.304788 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.304797 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.304804 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:33Z","lastTransitionTime":"2025-11-23T06:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.406459 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.406490 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.406499 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.406510 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.406518 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:33Z","lastTransitionTime":"2025-11-23T06:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.507971 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.508006 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.508033 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.508055 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.508064 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:33Z","lastTransitionTime":"2025-11-23T06:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.610549 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.610575 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.610585 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.610596 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.610603 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:33Z","lastTransitionTime":"2025-11-23T06:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.712494 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.712521 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.712531 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.712539 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.712546 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:33Z","lastTransitionTime":"2025-11-23T06:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.815024 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.815083 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.815092 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.815105 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.815113 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:33Z","lastTransitionTime":"2025-11-23T06:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.916252 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.916285 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.916292 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.916304 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:33 crc kubenswrapper[4559]: I1123 06:46:33.916311 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:33Z","lastTransitionTime":"2025-11-23T06:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.017934 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.017962 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.017970 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.017981 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.017990 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:34Z","lastTransitionTime":"2025-11-23T06:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.119618 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.119656 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.119665 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.119678 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.119691 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:34Z","lastTransitionTime":"2025-11-23T06:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.221370 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.221404 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.221413 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.221425 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.221437 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:34Z","lastTransitionTime":"2025-11-23T06:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.272962 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:34 crc kubenswrapper[4559]: E1123 06:46:34.273071 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.323132 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.323163 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.323170 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.323178 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.323186 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:34Z","lastTransitionTime":"2025-11-23T06:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.424846 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.424874 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.424882 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.424891 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.424898 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:34Z","lastTransitionTime":"2025-11-23T06:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.526863 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.526893 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.526918 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.526930 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.526938 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:34Z","lastTransitionTime":"2025-11-23T06:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.628463 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.628513 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.628522 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.628533 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.628542 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:34Z","lastTransitionTime":"2025-11-23T06:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.730443 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.730472 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.730480 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.730492 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.730500 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:34Z","lastTransitionTime":"2025-11-23T06:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.832305 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.832330 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.832340 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.832351 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.832357 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:34Z","lastTransitionTime":"2025-11-23T06:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.934262 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.934304 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.934314 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.934329 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:34 crc kubenswrapper[4559]: I1123 06:46:34.934343 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:34Z","lastTransitionTime":"2025-11-23T06:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.037317 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.037348 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.037357 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.037383 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.037392 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:35Z","lastTransitionTime":"2025-11-23T06:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.139139 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.139358 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.139416 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.139483 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.139546 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:35Z","lastTransitionTime":"2025-11-23T06:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.241297 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.241334 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.241343 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.241357 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.241364 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:35Z","lastTransitionTime":"2025-11-23T06:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.273567 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.273610 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.273580 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:35 crc kubenswrapper[4559]: E1123 06:46:35.273706 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:35 crc kubenswrapper[4559]: E1123 06:46:35.273783 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:35 crc kubenswrapper[4559]: E1123 06:46:35.273919 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.342818 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.342841 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.342850 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.342862 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.342871 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:35Z","lastTransitionTime":"2025-11-23T06:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.444632 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.444698 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.444707 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.444719 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.444727 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:35Z","lastTransitionTime":"2025-11-23T06:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.546586 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.546615 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.546624 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.546638 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.546663 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:35Z","lastTransitionTime":"2025-11-23T06:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.648502 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.648531 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.648539 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.648548 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.648556 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:35Z","lastTransitionTime":"2025-11-23T06:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.750854 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.750886 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.750895 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.750909 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.750918 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:35Z","lastTransitionTime":"2025-11-23T06:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.852597 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.852628 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.852636 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.852671 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.852681 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:35Z","lastTransitionTime":"2025-11-23T06:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.953980 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.954039 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.954050 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.954078 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:35 crc kubenswrapper[4559]: I1123 06:46:35.954086 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:35Z","lastTransitionTime":"2025-11-23T06:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.056024 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.056064 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.056072 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.056081 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.056091 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:36Z","lastTransitionTime":"2025-11-23T06:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.157349 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.157370 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.157430 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.157460 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.157478 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:36Z","lastTransitionTime":"2025-11-23T06:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.259713 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.259738 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.259746 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.259758 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.259766 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:36Z","lastTransitionTime":"2025-11-23T06:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.273061 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:36 crc kubenswrapper[4559]: E1123 06:46:36.273151 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.281764 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.361805 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.361829 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.361837 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.361847 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.361855 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:36Z","lastTransitionTime":"2025-11-23T06:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.463704 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.463728 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.463736 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.463746 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.463755 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:36Z","lastTransitionTime":"2025-11-23T06:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.565318 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.565361 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.565370 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.565383 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.565392 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:36Z","lastTransitionTime":"2025-11-23T06:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.667515 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.667545 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.667553 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.667565 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.667574 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:36Z","lastTransitionTime":"2025-11-23T06:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.769109 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.769139 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.769148 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.769158 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.769166 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:36Z","lastTransitionTime":"2025-11-23T06:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.870714 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.870749 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.870759 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.870771 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.870781 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:36Z","lastTransitionTime":"2025-11-23T06:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.972628 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.972679 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.972688 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.972701 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:36 crc kubenswrapper[4559]: I1123 06:46:36.972710 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:36Z","lastTransitionTime":"2025-11-23T06:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.073863 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.073895 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.073905 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.073916 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.073926 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:37Z","lastTransitionTime":"2025-11-23T06:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.175828 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.175851 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.175859 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.175869 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.175876 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:37Z","lastTransitionTime":"2025-11-23T06:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.273119 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.273176 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:37 crc kubenswrapper[4559]: E1123 06:46:37.273200 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.273259 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:37 crc kubenswrapper[4559]: E1123 06:46:37.273298 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:37 crc kubenswrapper[4559]: E1123 06:46:37.273374 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.277296 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.277324 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.277332 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.277363 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.277372 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:37Z","lastTransitionTime":"2025-11-23T06:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.379325 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.379351 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.379359 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.379369 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.379376 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:37Z","lastTransitionTime":"2025-11-23T06:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.480935 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.480962 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.480969 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.480979 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.480988 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:37Z","lastTransitionTime":"2025-11-23T06:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.582433 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.582459 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.582469 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.582480 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.582490 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:37Z","lastTransitionTime":"2025-11-23T06:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.616952 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/abb6827e-709d-4ea9-8090-e4f541b1ff3f-metrics-certs\") pod \"network-metrics-daemon-svb6z\" (UID: \"abb6827e-709d-4ea9-8090-e4f541b1ff3f\") " pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:37 crc kubenswrapper[4559]: E1123 06:46:37.617049 4559 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:46:37 crc kubenswrapper[4559]: E1123 06:46:37.617083 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/abb6827e-709d-4ea9-8090-e4f541b1ff3f-metrics-certs podName:abb6827e-709d-4ea9-8090-e4f541b1ff3f nodeName:}" failed. No retries permitted until 2025-11-23 06:47:41.617072377 +0000 UTC m=+163.639057991 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/abb6827e-709d-4ea9-8090-e4f541b1ff3f-metrics-certs") pod "network-metrics-daemon-svb6z" (UID: "abb6827e-709d-4ea9-8090-e4f541b1ff3f") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.684639 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.684677 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.684690 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.684700 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.684707 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:37Z","lastTransitionTime":"2025-11-23T06:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.786436 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.786468 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.786479 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.786504 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.786512 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:37Z","lastTransitionTime":"2025-11-23T06:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.888482 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.888506 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.888513 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.888522 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.888530 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:37Z","lastTransitionTime":"2025-11-23T06:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.990336 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.990364 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.990372 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.990381 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:37 crc kubenswrapper[4559]: I1123 06:46:37.990388 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:37Z","lastTransitionTime":"2025-11-23T06:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.092235 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.092275 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.092285 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.092298 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.092308 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:38Z","lastTransitionTime":"2025-11-23T06:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.193451 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.193470 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.193478 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.193486 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.193493 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:38Z","lastTransitionTime":"2025-11-23T06:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.272843 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:38 crc kubenswrapper[4559]: E1123 06:46:38.272928 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.294653 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.294793 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.294885 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.294971 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.295063 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:38Z","lastTransitionTime":"2025-11-23T06:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.304620 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-w5tlv" podStartSLOduration=79.304608476 podStartE2EDuration="1m19.304608476s" podCreationTimestamp="2025-11-23 06:45:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:46:38.295900232 +0000 UTC m=+100.317885847" watchObservedRunningTime="2025-11-23 06:46:38.304608476 +0000 UTC m=+100.326594100" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.304750 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=45.304746584 podStartE2EDuration="45.304746584s" podCreationTimestamp="2025-11-23 06:45:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:46:38.304400415 +0000 UTC m=+100.326386039" watchObservedRunningTime="2025-11-23 06:46:38.304746584 +0000 UTC m=+100.326732199" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.325635 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=78.325623481 podStartE2EDuration="1m18.325623481s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:46:38.32541497 +0000 UTC m=+100.347400584" watchObservedRunningTime="2025-11-23 06:46:38.325623481 +0000 UTC m=+100.347609095" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.352281 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-qn4h8" podStartSLOduration=78.352272693 podStartE2EDuration="1m18.352272693s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:46:38.352130957 +0000 UTC m=+100.374116581" watchObservedRunningTime="2025-11-23 06:46:38.352272693 +0000 UTC m=+100.374258307" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.362582 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-rlxnw" podStartSLOduration=78.362573734 podStartE2EDuration="1m18.362573734s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:46:38.362327402 +0000 UTC m=+100.384313016" watchObservedRunningTime="2025-11-23 06:46:38.362573734 +0000 UTC m=+100.384559337" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.387125 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=82.387115103 podStartE2EDuration="1m22.387115103s" podCreationTimestamp="2025-11-23 06:45:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:46:38.386341201 +0000 UTC m=+100.408326804" watchObservedRunningTime="2025-11-23 06:46:38.387115103 +0000 UTC m=+100.409100717" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.387428 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=11.387424463 podStartE2EDuration="11.387424463s" podCreationTimestamp="2025-11-23 06:46:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:46:38.375952856 +0000 UTC m=+100.397938470" watchObservedRunningTime="2025-11-23 06:46:38.387424463 +0000 UTC m=+100.409410077" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.396493 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.396516 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.396525 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.396536 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.396544 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:38Z","lastTransitionTime":"2025-11-23T06:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.408514 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podStartSLOduration=78.408500843 podStartE2EDuration="1m18.408500843s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:46:38.393526157 +0000 UTC m=+100.415511772" watchObservedRunningTime="2025-11-23 06:46:38.408500843 +0000 UTC m=+100.430486447" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.427683 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=2.427673131 podStartE2EDuration="2.427673131s" podCreationTimestamp="2025-11-23 06:46:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:46:38.427193792 +0000 UTC m=+100.449179407" watchObservedRunningTime="2025-11-23 06:46:38.427673131 +0000 UTC m=+100.449658745" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.450906 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9k8xh" podStartSLOduration=78.450890307 podStartE2EDuration="1m18.450890307s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:46:38.442124436 +0000 UTC m=+100.464110050" watchObservedRunningTime="2025-11-23 06:46:38.450890307 +0000 UTC m=+100.472875921" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.457583 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-njqzh" podStartSLOduration=79.457572289 podStartE2EDuration="1m19.457572289s" podCreationTimestamp="2025-11-23 06:45:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:46:38.456867047 +0000 UTC m=+100.478852661" watchObservedRunningTime="2025-11-23 06:46:38.457572289 +0000 UTC m=+100.479557894" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.498321 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.498350 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.498358 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.498369 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.498377 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:38Z","lastTransitionTime":"2025-11-23T06:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.599582 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.599627 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.599637 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.599681 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.599690 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:38Z","lastTransitionTime":"2025-11-23T06:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.701337 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.701364 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.701374 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.701386 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.701415 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:38Z","lastTransitionTime":"2025-11-23T06:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.803299 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.803333 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.803343 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.803357 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.803366 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:38Z","lastTransitionTime":"2025-11-23T06:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.905073 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.905097 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.905106 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.905116 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:38 crc kubenswrapper[4559]: I1123 06:46:38.905141 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:38Z","lastTransitionTime":"2025-11-23T06:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.006639 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.006682 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.006690 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.006699 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.006708 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:39Z","lastTransitionTime":"2025-11-23T06:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.108431 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.108455 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.108462 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.108496 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.108504 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:39Z","lastTransitionTime":"2025-11-23T06:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.210004 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.210045 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.210053 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.210080 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.210088 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:39Z","lastTransitionTime":"2025-11-23T06:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.273036 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.273066 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:39 crc kubenswrapper[4559]: E1123 06:46:39.273123 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.273182 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:39 crc kubenswrapper[4559]: E1123 06:46:39.273265 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:39 crc kubenswrapper[4559]: E1123 06:46:39.273502 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.311943 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.311965 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.311972 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.311983 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.311990 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:39Z","lastTransitionTime":"2025-11-23T06:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.413988 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.414134 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.414212 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.414289 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.414355 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:39Z","lastTransitionTime":"2025-11-23T06:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.515768 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.515899 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.515987 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.516077 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.516143 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:39Z","lastTransitionTime":"2025-11-23T06:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.618467 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.618495 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.618503 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.618514 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.618522 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:39Z","lastTransitionTime":"2025-11-23T06:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.720217 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.720241 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.720248 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.720258 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.720265 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:39Z","lastTransitionTime":"2025-11-23T06:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.822041 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.822067 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.822075 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.822084 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.822091 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:39Z","lastTransitionTime":"2025-11-23T06:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.923125 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.923153 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.923161 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.923170 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:39 crc kubenswrapper[4559]: I1123 06:46:39.923177 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:39Z","lastTransitionTime":"2025-11-23T06:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.024574 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.024610 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.024619 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.024633 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.024656 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:40Z","lastTransitionTime":"2025-11-23T06:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.126059 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.126094 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.126108 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.126122 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.126130 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:40Z","lastTransitionTime":"2025-11-23T06:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.228422 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.228445 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.228471 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.228482 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.228489 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:40Z","lastTransitionTime":"2025-11-23T06:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.273270 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:40 crc kubenswrapper[4559]: E1123 06:46:40.273376 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.330191 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.330217 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.330225 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.330237 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.330246 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:40Z","lastTransitionTime":"2025-11-23T06:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.431837 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.431865 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.431873 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.431884 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.431892 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:40Z","lastTransitionTime":"2025-11-23T06:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.533423 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.533457 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.533465 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.533478 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.533487 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:40Z","lastTransitionTime":"2025-11-23T06:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.635325 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.635385 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.635396 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.635413 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.635422 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:40Z","lastTransitionTime":"2025-11-23T06:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.643729 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.643755 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.643764 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.643773 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.643779 4559 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T06:46:40Z","lastTransitionTime":"2025-11-23T06:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.669105 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-bj5hn"] Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.669360 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bj5hn" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.670519 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.670689 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.670913 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.671033 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.741184 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6a290521-6ca4-4d84-b487-7de2883f43fa-service-ca\") pod \"cluster-version-operator-5c965bbfc6-bj5hn\" (UID: \"6a290521-6ca4-4d84-b487-7de2883f43fa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bj5hn" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.741233 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6a290521-6ca4-4d84-b487-7de2883f43fa-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-bj5hn\" (UID: \"6a290521-6ca4-4d84-b487-7de2883f43fa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bj5hn" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.741254 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/6a290521-6ca4-4d84-b487-7de2883f43fa-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-bj5hn\" (UID: \"6a290521-6ca4-4d84-b487-7de2883f43fa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bj5hn" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.741309 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/6a290521-6ca4-4d84-b487-7de2883f43fa-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-bj5hn\" (UID: \"6a290521-6ca4-4d84-b487-7de2883f43fa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bj5hn" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.741352 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a290521-6ca4-4d84-b487-7de2883f43fa-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-bj5hn\" (UID: \"6a290521-6ca4-4d84-b487-7de2883f43fa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bj5hn" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.842439 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6a290521-6ca4-4d84-b487-7de2883f43fa-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-bj5hn\" (UID: \"6a290521-6ca4-4d84-b487-7de2883f43fa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bj5hn" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.842474 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/6a290521-6ca4-4d84-b487-7de2883f43fa-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-bj5hn\" (UID: \"6a290521-6ca4-4d84-b487-7de2883f43fa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bj5hn" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.842498 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/6a290521-6ca4-4d84-b487-7de2883f43fa-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-bj5hn\" (UID: \"6a290521-6ca4-4d84-b487-7de2883f43fa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bj5hn" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.842535 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a290521-6ca4-4d84-b487-7de2883f43fa-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-bj5hn\" (UID: \"6a290521-6ca4-4d84-b487-7de2883f43fa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bj5hn" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.842558 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6a290521-6ca4-4d84-b487-7de2883f43fa-service-ca\") pod \"cluster-version-operator-5c965bbfc6-bj5hn\" (UID: \"6a290521-6ca4-4d84-b487-7de2883f43fa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bj5hn" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.842596 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/6a290521-6ca4-4d84-b487-7de2883f43fa-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-bj5hn\" (UID: \"6a290521-6ca4-4d84-b487-7de2883f43fa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bj5hn" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.842609 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/6a290521-6ca4-4d84-b487-7de2883f43fa-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-bj5hn\" (UID: \"6a290521-6ca4-4d84-b487-7de2883f43fa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bj5hn" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.843245 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6a290521-6ca4-4d84-b487-7de2883f43fa-service-ca\") pod \"cluster-version-operator-5c965bbfc6-bj5hn\" (UID: \"6a290521-6ca4-4d84-b487-7de2883f43fa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bj5hn" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.846929 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a290521-6ca4-4d84-b487-7de2883f43fa-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-bj5hn\" (UID: \"6a290521-6ca4-4d84-b487-7de2883f43fa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bj5hn" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.855413 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6a290521-6ca4-4d84-b487-7de2883f43fa-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-bj5hn\" (UID: \"6a290521-6ca4-4d84-b487-7de2883f43fa\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bj5hn" Nov 23 06:46:40 crc kubenswrapper[4559]: I1123 06:46:40.979207 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bj5hn" Nov 23 06:46:40 crc kubenswrapper[4559]: W1123 06:46:40.988625 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a290521_6ca4_4d84_b487_7de2883f43fa.slice/crio-72fc19f70a4f8fb2ecfeaabaf329056e753faab8402d06a1aa499a10f9874b77 WatchSource:0}: Error finding container 72fc19f70a4f8fb2ecfeaabaf329056e753faab8402d06a1aa499a10f9874b77: Status 404 returned error can't find the container with id 72fc19f70a4f8fb2ecfeaabaf329056e753faab8402d06a1aa499a10f9874b77 Nov 23 06:46:41 crc kubenswrapper[4559]: I1123 06:46:41.272838 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:41 crc kubenswrapper[4559]: I1123 06:46:41.272902 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:41 crc kubenswrapper[4559]: E1123 06:46:41.272992 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:41 crc kubenswrapper[4559]: I1123 06:46:41.273019 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:41 crc kubenswrapper[4559]: E1123 06:46:41.273075 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:41 crc kubenswrapper[4559]: E1123 06:46:41.273145 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:41 crc kubenswrapper[4559]: I1123 06:46:41.591710 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bj5hn" event={"ID":"6a290521-6ca4-4d84-b487-7de2883f43fa","Type":"ContainerStarted","Data":"979aa412e61f8d8d8ca751ec8a800d7441720d4f9ef649f525f57542777464e3"} Nov 23 06:46:41 crc kubenswrapper[4559]: I1123 06:46:41.591770 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bj5hn" event={"ID":"6a290521-6ca4-4d84-b487-7de2883f43fa","Type":"ContainerStarted","Data":"72fc19f70a4f8fb2ecfeaabaf329056e753faab8402d06a1aa499a10f9874b77"} Nov 23 06:46:41 crc kubenswrapper[4559]: I1123 06:46:41.601435 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bj5hn" podStartSLOduration=81.60141979 podStartE2EDuration="1m21.60141979s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:46:41.601197673 +0000 UTC m=+103.623183288" watchObservedRunningTime="2025-11-23 06:46:41.60141979 +0000 UTC m=+103.623405405" Nov 23 06:46:42 crc kubenswrapper[4559]: I1123 06:46:42.273177 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:42 crc kubenswrapper[4559]: E1123 06:46:42.273288 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:43 crc kubenswrapper[4559]: I1123 06:46:43.273069 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:43 crc kubenswrapper[4559]: E1123 06:46:43.273378 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:43 crc kubenswrapper[4559]: I1123 06:46:43.273113 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:43 crc kubenswrapper[4559]: E1123 06:46:43.273896 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:43 crc kubenswrapper[4559]: I1123 06:46:43.273670 4559 scope.go:117] "RemoveContainer" containerID="28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da" Nov 23 06:46:43 crc kubenswrapper[4559]: I1123 06:46:43.273077 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:43 crc kubenswrapper[4559]: E1123 06:46:43.274211 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:43 crc kubenswrapper[4559]: E1123 06:46:43.274426 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jt56f_openshift-ovn-kubernetes(b12e257e-7887-4795-9221-4db8fd5856dd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" Nov 23 06:46:44 crc kubenswrapper[4559]: I1123 06:46:44.273073 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:44 crc kubenswrapper[4559]: E1123 06:46:44.273378 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:45 crc kubenswrapper[4559]: I1123 06:46:45.272854 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:45 crc kubenswrapper[4559]: I1123 06:46:45.272879 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:45 crc kubenswrapper[4559]: I1123 06:46:45.272880 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:45 crc kubenswrapper[4559]: E1123 06:46:45.272956 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:45 crc kubenswrapper[4559]: E1123 06:46:45.273052 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:45 crc kubenswrapper[4559]: E1123 06:46:45.273324 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:46 crc kubenswrapper[4559]: I1123 06:46:46.273086 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:46 crc kubenswrapper[4559]: E1123 06:46:46.273304 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:47 crc kubenswrapper[4559]: I1123 06:46:47.272782 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:47 crc kubenswrapper[4559]: I1123 06:46:47.272884 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:47 crc kubenswrapper[4559]: I1123 06:46:47.272972 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:47 crc kubenswrapper[4559]: E1123 06:46:47.272964 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:47 crc kubenswrapper[4559]: E1123 06:46:47.273105 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:47 crc kubenswrapper[4559]: E1123 06:46:47.273182 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:48 crc kubenswrapper[4559]: I1123 06:46:48.273065 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:48 crc kubenswrapper[4559]: E1123 06:46:48.273916 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:49 crc kubenswrapper[4559]: I1123 06:46:49.272846 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:49 crc kubenswrapper[4559]: I1123 06:46:49.272854 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:49 crc kubenswrapper[4559]: E1123 06:46:49.273096 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:49 crc kubenswrapper[4559]: E1123 06:46:49.273199 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:49 crc kubenswrapper[4559]: I1123 06:46:49.273266 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:49 crc kubenswrapper[4559]: E1123 06:46:49.273314 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:50 crc kubenswrapper[4559]: I1123 06:46:50.272917 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:50 crc kubenswrapper[4559]: E1123 06:46:50.273196 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:51 crc kubenswrapper[4559]: I1123 06:46:51.273269 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:51 crc kubenswrapper[4559]: I1123 06:46:51.273339 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:51 crc kubenswrapper[4559]: E1123 06:46:51.273347 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:51 crc kubenswrapper[4559]: I1123 06:46:51.273373 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:51 crc kubenswrapper[4559]: E1123 06:46:51.273391 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:51 crc kubenswrapper[4559]: E1123 06:46:51.273443 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:52 crc kubenswrapper[4559]: I1123 06:46:52.273476 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:52 crc kubenswrapper[4559]: E1123 06:46:52.273572 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:53 crc kubenswrapper[4559]: I1123 06:46:53.272484 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:53 crc kubenswrapper[4559]: E1123 06:46:53.272747 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:53 crc kubenswrapper[4559]: I1123 06:46:53.272536 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:53 crc kubenswrapper[4559]: I1123 06:46:53.272484 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:53 crc kubenswrapper[4559]: E1123 06:46:53.272803 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:53 crc kubenswrapper[4559]: E1123 06:46:53.272858 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:53 crc kubenswrapper[4559]: I1123 06:46:53.617895 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qn4h8_18091d80-4837-4894-8583-9a2bd30e10b9/kube-multus/1.log" Nov 23 06:46:53 crc kubenswrapper[4559]: I1123 06:46:53.618475 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qn4h8_18091d80-4837-4894-8583-9a2bd30e10b9/kube-multus/0.log" Nov 23 06:46:53 crc kubenswrapper[4559]: I1123 06:46:53.618571 4559 generic.go:334] "Generic (PLEG): container finished" podID="18091d80-4837-4894-8583-9a2bd30e10b9" containerID="73ab4ccc28f84b0b246d2a4464c101f4f3113b2e910eea31f0effd6b89a0a71f" exitCode=1 Nov 23 06:46:53 crc kubenswrapper[4559]: I1123 06:46:53.618599 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qn4h8" event={"ID":"18091d80-4837-4894-8583-9a2bd30e10b9","Type":"ContainerDied","Data":"73ab4ccc28f84b0b246d2a4464c101f4f3113b2e910eea31f0effd6b89a0a71f"} Nov 23 06:46:53 crc kubenswrapper[4559]: I1123 06:46:53.618759 4559 scope.go:117] "RemoveContainer" containerID="aa2b86b006c0ebb7b03c7e7db7e405a3e781fc48e0570322152bb4a5f8c1ff64" Nov 23 06:46:53 crc kubenswrapper[4559]: I1123 06:46:53.619068 4559 scope.go:117] "RemoveContainer" containerID="73ab4ccc28f84b0b246d2a4464c101f4f3113b2e910eea31f0effd6b89a0a71f" Nov 23 06:46:53 crc kubenswrapper[4559]: E1123 06:46:53.619246 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-qn4h8_openshift-multus(18091d80-4837-4894-8583-9a2bd30e10b9)\"" pod="openshift-multus/multus-qn4h8" podUID="18091d80-4837-4894-8583-9a2bd30e10b9" Nov 23 06:46:54 crc kubenswrapper[4559]: I1123 06:46:54.273236 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:54 crc kubenswrapper[4559]: E1123 06:46:54.273327 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:54 crc kubenswrapper[4559]: I1123 06:46:54.622280 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qn4h8_18091d80-4837-4894-8583-9a2bd30e10b9/kube-multus/1.log" Nov 23 06:46:55 crc kubenswrapper[4559]: I1123 06:46:55.273177 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:55 crc kubenswrapper[4559]: I1123 06:46:55.273204 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:55 crc kubenswrapper[4559]: I1123 06:46:55.273209 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:55 crc kubenswrapper[4559]: E1123 06:46:55.273263 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:55 crc kubenswrapper[4559]: E1123 06:46:55.273321 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:55 crc kubenswrapper[4559]: E1123 06:46:55.273381 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:56 crc kubenswrapper[4559]: I1123 06:46:56.273276 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:56 crc kubenswrapper[4559]: E1123 06:46:56.273364 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:57 crc kubenswrapper[4559]: I1123 06:46:57.272526 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:57 crc kubenswrapper[4559]: I1123 06:46:57.272572 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:57 crc kubenswrapper[4559]: I1123 06:46:57.272571 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:57 crc kubenswrapper[4559]: E1123 06:46:57.272637 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:57 crc kubenswrapper[4559]: E1123 06:46:57.272697 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:46:57 crc kubenswrapper[4559]: E1123 06:46:57.272765 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:58 crc kubenswrapper[4559]: I1123 06:46:58.272748 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:46:58 crc kubenswrapper[4559]: E1123 06:46:58.273505 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:46:58 crc kubenswrapper[4559]: I1123 06:46:58.273932 4559 scope.go:117] "RemoveContainer" containerID="28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da" Nov 23 06:46:58 crc kubenswrapper[4559]: E1123 06:46:58.344814 4559 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 23 06:46:58 crc kubenswrapper[4559]: I1123 06:46:58.632788 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jt56f_b12e257e-7887-4795-9221-4db8fd5856dd/ovnkube-controller/3.log" Nov 23 06:46:58 crc kubenswrapper[4559]: I1123 06:46:58.634739 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" event={"ID":"b12e257e-7887-4795-9221-4db8fd5856dd","Type":"ContainerStarted","Data":"0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8"} Nov 23 06:46:58 crc kubenswrapper[4559]: I1123 06:46:58.635044 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:46:58 crc kubenswrapper[4559]: I1123 06:46:58.653467 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" podStartSLOduration=98.653454883 podStartE2EDuration="1m38.653454883s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:46:58.651956863 +0000 UTC m=+120.673942477" watchObservedRunningTime="2025-11-23 06:46:58.653454883 +0000 UTC m=+120.675440497" Nov 23 06:46:58 crc kubenswrapper[4559]: I1123 06:46:58.843981 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-svb6z"] Nov 23 06:46:58 crc kubenswrapper[4559]: I1123 06:46:58.844079 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:46:58 crc kubenswrapper[4559]: E1123 06:46:58.844153 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:46:59 crc kubenswrapper[4559]: I1123 06:46:59.273076 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:46:59 crc kubenswrapper[4559]: E1123 06:46:59.273329 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:46:59 crc kubenswrapper[4559]: I1123 06:46:59.273076 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:46:59 crc kubenswrapper[4559]: E1123 06:46:59.273427 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:47:00 crc kubenswrapper[4559]: I1123 06:47:00.273041 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:47:00 crc kubenswrapper[4559]: E1123 06:47:00.273137 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:47:00 crc kubenswrapper[4559]: I1123 06:47:00.273041 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:47:00 crc kubenswrapper[4559]: E1123 06:47:00.273238 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:47:01 crc kubenswrapper[4559]: I1123 06:47:01.272961 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:47:01 crc kubenswrapper[4559]: I1123 06:47:01.272991 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:47:01 crc kubenswrapper[4559]: E1123 06:47:01.273080 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:47:01 crc kubenswrapper[4559]: E1123 06:47:01.273181 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:47:02 crc kubenswrapper[4559]: I1123 06:47:02.273415 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:47:02 crc kubenswrapper[4559]: E1123 06:47:02.273922 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:47:02 crc kubenswrapper[4559]: I1123 06:47:02.274099 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:47:02 crc kubenswrapper[4559]: E1123 06:47:02.274209 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:47:03 crc kubenswrapper[4559]: I1123 06:47:03.273211 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:47:03 crc kubenswrapper[4559]: I1123 06:47:03.273278 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:47:03 crc kubenswrapper[4559]: E1123 06:47:03.273315 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:47:03 crc kubenswrapper[4559]: E1123 06:47:03.273383 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:47:03 crc kubenswrapper[4559]: E1123 06:47:03.335285 4559 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 23 06:47:04 crc kubenswrapper[4559]: I1123 06:47:04.273344 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:47:04 crc kubenswrapper[4559]: E1123 06:47:04.273446 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:47:04 crc kubenswrapper[4559]: I1123 06:47:04.273922 4559 scope.go:117] "RemoveContainer" containerID="73ab4ccc28f84b0b246d2a4464c101f4f3113b2e910eea31f0effd6b89a0a71f" Nov 23 06:47:04 crc kubenswrapper[4559]: I1123 06:47:04.274017 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:47:04 crc kubenswrapper[4559]: E1123 06:47:04.274117 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:47:04 crc kubenswrapper[4559]: I1123 06:47:04.649326 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qn4h8_18091d80-4837-4894-8583-9a2bd30e10b9/kube-multus/1.log" Nov 23 06:47:04 crc kubenswrapper[4559]: I1123 06:47:04.649519 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qn4h8" event={"ID":"18091d80-4837-4894-8583-9a2bd30e10b9","Type":"ContainerStarted","Data":"43e22f4ac21e7304b67c5d85e3daca35952d0af501662a9f20bff79f4b065ff7"} Nov 23 06:47:05 crc kubenswrapper[4559]: I1123 06:47:05.273164 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:47:05 crc kubenswrapper[4559]: I1123 06:47:05.273168 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:47:05 crc kubenswrapper[4559]: E1123 06:47:05.273376 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:47:05 crc kubenswrapper[4559]: E1123 06:47:05.273269 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:47:06 crc kubenswrapper[4559]: I1123 06:47:06.272701 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:47:06 crc kubenswrapper[4559]: I1123 06:47:06.272724 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:47:06 crc kubenswrapper[4559]: E1123 06:47:06.272806 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:47:06 crc kubenswrapper[4559]: E1123 06:47:06.272881 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:47:07 crc kubenswrapper[4559]: I1123 06:47:07.273069 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:47:07 crc kubenswrapper[4559]: I1123 06:47:07.273069 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:47:07 crc kubenswrapper[4559]: E1123 06:47:07.273235 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 06:47:07 crc kubenswrapper[4559]: E1123 06:47:07.273172 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 06:47:08 crc kubenswrapper[4559]: I1123 06:47:08.272679 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:47:08 crc kubenswrapper[4559]: I1123 06:47:08.272757 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:47:08 crc kubenswrapper[4559]: E1123 06:47:08.273403 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 06:47:08 crc kubenswrapper[4559]: E1123 06:47:08.273535 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-svb6z" podUID="abb6827e-709d-4ea9-8090-e4f541b1ff3f" Nov 23 06:47:09 crc kubenswrapper[4559]: I1123 06:47:09.272745 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:47:09 crc kubenswrapper[4559]: I1123 06:47:09.272796 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:47:09 crc kubenswrapper[4559]: I1123 06:47:09.274017 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 23 06:47:09 crc kubenswrapper[4559]: I1123 06:47:09.274195 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 23 06:47:10 crc kubenswrapper[4559]: I1123 06:47:10.273129 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:47:10 crc kubenswrapper[4559]: I1123 06:47:10.273207 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:47:10 crc kubenswrapper[4559]: I1123 06:47:10.275289 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 23 06:47:10 crc kubenswrapper[4559]: I1123 06:47:10.275358 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 23 06:47:10 crc kubenswrapper[4559]: I1123 06:47:10.275372 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 23 06:47:10 crc kubenswrapper[4559]: I1123 06:47:10.275532 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.189603 4559 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.211571 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-ztnsp"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.212009 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-ztnsp" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.212210 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-5kfl6"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.212723 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.212724 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r5ts6"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.213147 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-r5ts6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.213731 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.217143 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.217311 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.217514 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.217566 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.217587 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.217696 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.217856 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.217977 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.218147 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.218239 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.218296 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.218567 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.218588 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.218721 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.218888 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.218907 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.218941 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.218914 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.225701 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.229249 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.229272 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.232898 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.233259 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.233573 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q79mk"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.233967 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q79mk" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.234227 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-q5d8b"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.234500 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-q5d8b" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.234705 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-l7z49"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.235146 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7z49" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.235364 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-kg77n"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.235634 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.235928 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-xvd4d"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.236345 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.236536 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.236847 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.237175 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-fjv2r"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.237743 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fjv2r" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.237761 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j665r"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.238273 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j665r" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.238426 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-p9rps"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.238805 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-p9rps" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.239041 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.239052 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-rqftc"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.242371 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.242665 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.243380 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.243557 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.245437 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.245615 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8llfp"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.245874 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.245976 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-rqftc" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.250431 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.250535 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.250555 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.250631 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.250672 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.250829 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.250840 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.250827 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-dlj2h"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.250961 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.251068 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.251065 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8llfp" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.251165 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.251246 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-ztnsp"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.251253 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.251282 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-dlj2h" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.251396 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.251422 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.251578 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.251583 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.251776 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.251888 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.251783 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.251801 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.252131 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.251970 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.252567 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.252736 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.252828 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.252940 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.253579 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.253725 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.253863 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.254090 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.254258 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.254357 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.254508 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.254521 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-9m2sk"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.254668 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.254995 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-9m2sk" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.257953 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ggt5f"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.258300 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.258566 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.258764 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.258911 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.260102 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.260223 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.260332 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.260416 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.260522 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.260540 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.260636 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.260807 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.260960 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.261018 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.261089 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.261132 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.261228 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.261246 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.261322 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.261750 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.261890 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.262009 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.261899 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.262197 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.263017 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.263266 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.263359 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.263407 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.263767 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.263855 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.265858 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-t2rfc"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.266393 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qt4fl"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.266605 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r779q"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.266957 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ggt5f" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.267255 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.267288 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t2rfc" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.267566 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qt4fl" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.267921 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8xfrw"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.268154 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-dnz8k"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.268741 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zr8cm"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.268968 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-9gwbh"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.269195 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-9gwbh" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.272065 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ccrq2"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.272465 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r779q" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.272467 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-8wt84"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.272616 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.272736 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8xfrw" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.272781 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dnz8k" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.272814 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zr8cm" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.273044 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-78zq2"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.273371 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-78zq2" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.273896 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-8wt84" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.274689 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwb8n\" (UniqueName: \"kubernetes.io/projected/b3e2ea56-3525-4e78-a7c8-289a9a2406e6-kube-api-access-bwb8n\") pod \"cluster-image-registry-operator-dc59b4c8b-8llfp\" (UID: \"b3e2ea56-3525-4e78-a7c8-289a9a2406e6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8llfp" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.274716 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/994432aa-8794-4818-94bd-ce28d97c7163-audit\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.274717 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xwb9l"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.274737 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/534e137b-1c8d-45fb-9c1b-6804913d309b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-r5ts6\" (UID: \"534e137b-1c8d-45fb-9c1b-6804913d309b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r5ts6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.274755 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/e417e794-f6b2-4607-b2ce-3a737e8770b4-available-featuregates\") pod \"openshift-config-operator-7777fb866f-fjv2r\" (UID: \"e417e794-f6b2-4607-b2ce-3a737e8770b4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fjv2r" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.274770 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnflg\" (UniqueName: \"kubernetes.io/projected/994432aa-8794-4818-94bd-ce28d97c7163-kube-api-access-qnflg\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.274784 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/410014cb-6ba0-4610-acc5-38f3db8fcae2-auth-proxy-config\") pod \"machine-approver-56656f9798-l7z49\" (UID: \"410014cb-6ba0-4610-acc5-38f3db8fcae2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7z49" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.274798 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a883895-9775-4c18-962e-be4df71043c8-config\") pod \"openshift-apiserver-operator-796bbdcf4f-q79mk\" (UID: \"3a883895-9775-4c18-962e-be4df71043c8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q79mk" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.274813 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/994432aa-8794-4818-94bd-ce28d97c7163-etcd-client\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.274828 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/410014cb-6ba0-4610-acc5-38f3db8fcae2-config\") pod \"machine-approver-56656f9798-l7z49\" (UID: \"410014cb-6ba0-4610-acc5-38f3db8fcae2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7z49" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.274844 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e9514f9-f14f-4409-8fef-b89a37de70a6-config\") pod \"etcd-operator-b45778765-9m2sk\" (UID: \"9e9514f9-f14f-4409-8fef-b89a37de70a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9m2sk" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.274858 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/994432aa-8794-4818-94bd-ce28d97c7163-etcd-serving-ca\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.274881 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/994432aa-8794-4818-94bd-ce28d97c7163-serving-cert\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.274895 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6796a9bd-a428-4b0a-8723-d8a6a252de84-console-serving-cert\") pod \"console-f9d7485db-xvd4d\" (UID: \"6796a9bd-a428-4b0a-8723-d8a6a252de84\") " pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.274909 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6796a9bd-a428-4b0a-8723-d8a6a252de84-console-oauth-config\") pod \"console-f9d7485db-xvd4d\" (UID: \"6796a9bd-a428-4b0a-8723-d8a6a252de84\") " pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.274925 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e814799e-e374-4ad8-96a2-46e4d5defc2f-etcd-client\") pod \"apiserver-7bbb656c7d-ntvtt\" (UID: \"e814799e-e374-4ad8-96a2-46e4d5defc2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.274943 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b3e2ea56-3525-4e78-a7c8-289a9a2406e6-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-8llfp\" (UID: \"b3e2ea56-3525-4e78-a7c8-289a9a2406e6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8llfp" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.274959 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.274973 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b3e2ea56-3525-4e78-a7c8-289a9a2406e6-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-8llfp\" (UID: \"b3e2ea56-3525-4e78-a7c8-289a9a2406e6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8llfp" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.274987 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.275018 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08b0652b-f30f-4c2c-bd1a-75b80a66b144-config\") pod \"route-controller-manager-6576b87f9c-hzh5b\" (UID: \"08b0652b-f30f-4c2c-bd1a-75b80a66b144\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.275036 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f90213c-484b-4f9d-b064-363887589530-config\") pod \"authentication-operator-69f744f599-q5d8b\" (UID: \"7f90213c-484b-4f9d-b064-363887589530\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q5d8b" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.275050 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/58f094f8-8c16-4929-ae94-e52241c5b7d1-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-j665r\" (UID: \"58f094f8-8c16-4929-ae94-e52241c5b7d1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j665r" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.275090 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f90213c-484b-4f9d-b064-363887589530-serving-cert\") pod \"authentication-operator-69f744f599-q5d8b\" (UID: \"7f90213c-484b-4f9d-b064-363887589530\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q5d8b" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.275112 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f90213c-484b-4f9d-b064-363887589530-service-ca-bundle\") pod \"authentication-operator-69f744f599-q5d8b\" (UID: \"7f90213c-484b-4f9d-b064-363887589530\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q5d8b" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.275145 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/08b0652b-f30f-4c2c-bd1a-75b80a66b144-client-ca\") pod \"route-controller-manager-6576b87f9c-hzh5b\" (UID: \"08b0652b-f30f-4c2c-bd1a-75b80a66b144\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.275162 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6796a9bd-a428-4b0a-8723-d8a6a252de84-service-ca\") pod \"console-f9d7485db-xvd4d\" (UID: \"6796a9bd-a428-4b0a-8723-d8a6a252de84\") " pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.275260 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xwb9l" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.275849 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l6xn\" (UniqueName: \"kubernetes.io/projected/e814799e-e374-4ad8-96a2-46e4d5defc2f-kube-api-access-6l6xn\") pod \"apiserver-7bbb656c7d-ntvtt\" (UID: \"e814799e-e374-4ad8-96a2-46e4d5defc2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.275900 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e7b5892a-8920-4cef-94a1-3736328e9a39-audit-policies\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.275920 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z57mp\" (UniqueName: \"kubernetes.io/projected/e7b5892a-8920-4cef-94a1-3736328e9a39-kube-api-access-z57mp\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.275936 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/994432aa-8794-4818-94bd-ce28d97c7163-config\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.275951 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/994432aa-8794-4818-94bd-ce28d97c7163-image-import-ca\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.275965 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.275982 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdt5m\" (UniqueName: \"kubernetes.io/projected/18f99d77-1c2d-412b-93ae-1d1fc52f24ab-kube-api-access-vdt5m\") pod \"machine-api-operator-5694c8668f-ztnsp\" (UID: \"18f99d77-1c2d-412b-93ae-1d1fc52f24ab\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ztnsp" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.275997 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9e9514f9-f14f-4409-8fef-b89a37de70a6-etcd-client\") pod \"etcd-operator-b45778765-9m2sk\" (UID: \"9e9514f9-f14f-4409-8fef-b89a37de70a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9m2sk" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276012 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6796a9bd-a428-4b0a-8723-d8a6a252de84-trusted-ca-bundle\") pod \"console-f9d7485db-xvd4d\" (UID: \"6796a9bd-a428-4b0a-8723-d8a6a252de84\") " pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276027 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e814799e-e374-4ad8-96a2-46e4d5defc2f-audit-dir\") pod \"apiserver-7bbb656c7d-ntvtt\" (UID: \"e814799e-e374-4ad8-96a2-46e4d5defc2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276042 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/994432aa-8794-4818-94bd-ce28d97c7163-trusted-ca-bundle\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276057 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fjff\" (UniqueName: \"kubernetes.io/projected/6796a9bd-a428-4b0a-8723-d8a6a252de84-kube-api-access-8fjff\") pod \"console-f9d7485db-xvd4d\" (UID: \"6796a9bd-a428-4b0a-8723-d8a6a252de84\") " pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276072 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/18f99d77-1c2d-412b-93ae-1d1fc52f24ab-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-ztnsp\" (UID: \"18f99d77-1c2d-412b-93ae-1d1fc52f24ab\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ztnsp" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276084 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8xcm\" (UniqueName: \"kubernetes.io/projected/9e9514f9-f14f-4409-8fef-b89a37de70a6-kube-api-access-f8xcm\") pod \"etcd-operator-b45778765-9m2sk\" (UID: \"9e9514f9-f14f-4409-8fef-b89a37de70a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9m2sk" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276107 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/41ae564b-da35-45eb-9a27-fbf23bacaf8f-trusted-ca\") pod \"console-operator-58897d9998-rqftc\" (UID: \"41ae564b-da35-45eb-9a27-fbf23bacaf8f\") " pod="openshift-console-operator/console-operator-58897d9998-rqftc" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276122 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e9514f9-f14f-4409-8fef-b89a37de70a6-serving-cert\") pod \"etcd-operator-b45778765-9m2sk\" (UID: \"9e9514f9-f14f-4409-8fef-b89a37de70a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9m2sk" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276138 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/994432aa-8794-4818-94bd-ce28d97c7163-encryption-config\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276150 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e814799e-e374-4ad8-96a2-46e4d5defc2f-serving-cert\") pod \"apiserver-7bbb656c7d-ntvtt\" (UID: \"e814799e-e374-4ad8-96a2-46e4d5defc2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276163 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f90213c-484b-4f9d-b064-363887589530-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-q5d8b\" (UID: \"7f90213c-484b-4f9d-b064-363887589530\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q5d8b" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276184 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndlbq\" (UniqueName: \"kubernetes.io/projected/e417e794-f6b2-4607-b2ce-3a737e8770b4-kube-api-access-ndlbq\") pod \"openshift-config-operator-7777fb866f-fjv2r\" (UID: \"e417e794-f6b2-4607-b2ce-3a737e8770b4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fjv2r" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276196 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6mxz\" (UniqueName: \"kubernetes.io/projected/41ae564b-da35-45eb-9a27-fbf23bacaf8f-kube-api-access-b6mxz\") pod \"console-operator-58897d9998-rqftc\" (UID: \"41ae564b-da35-45eb-9a27-fbf23bacaf8f\") " pod="openshift-console-operator/console-operator-58897d9998-rqftc" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276216 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/9e9514f9-f14f-4409-8fef-b89a37de70a6-etcd-ca\") pod \"etcd-operator-b45778765-9m2sk\" (UID: \"9e9514f9-f14f-4409-8fef-b89a37de70a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9m2sk" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276231 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b3e2ea56-3525-4e78-a7c8-289a9a2406e6-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-8llfp\" (UID: \"b3e2ea56-3525-4e78-a7c8-289a9a2406e6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8llfp" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276246 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276259 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276274 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/16e9fcf3-adc7-4af9-a625-125ff496cfcf-metrics-tls\") pod \"dns-operator-744455d44c-dlj2h\" (UID: \"16e9fcf3-adc7-4af9-a625-125ff496cfcf\") " pod="openshift-dns-operator/dns-operator-744455d44c-dlj2h" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276289 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/534e137b-1c8d-45fb-9c1b-6804913d309b-client-ca\") pod \"controller-manager-879f6c89f-r5ts6\" (UID: \"534e137b-1c8d-45fb-9c1b-6804913d309b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r5ts6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276309 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/534e137b-1c8d-45fb-9c1b-6804913d309b-serving-cert\") pod \"controller-manager-879f6c89f-r5ts6\" (UID: \"534e137b-1c8d-45fb-9c1b-6804913d309b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r5ts6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276322 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/410014cb-6ba0-4610-acc5-38f3db8fcae2-machine-approver-tls\") pod \"machine-approver-56656f9798-l7z49\" (UID: \"410014cb-6ba0-4610-acc5-38f3db8fcae2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7z49" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276335 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e814799e-e374-4ad8-96a2-46e4d5defc2f-audit-policies\") pod \"apiserver-7bbb656c7d-ntvtt\" (UID: \"e814799e-e374-4ad8-96a2-46e4d5defc2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276347 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276368 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e417e794-f6b2-4607-b2ce-3a737e8770b4-serving-cert\") pod \"openshift-config-operator-7777fb866f-fjv2r\" (UID: \"e417e794-f6b2-4607-b2ce-3a737e8770b4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fjv2r" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276380 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfjqs\" (UniqueName: \"kubernetes.io/projected/3a883895-9775-4c18-962e-be4df71043c8-kube-api-access-sfjqs\") pod \"openshift-apiserver-operator-796bbdcf4f-q79mk\" (UID: \"3a883895-9775-4c18-962e-be4df71043c8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q79mk" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276392 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8hn2\" (UniqueName: \"kubernetes.io/projected/7f90213c-484b-4f9d-b064-363887589530-kube-api-access-g8hn2\") pod \"authentication-operator-69f744f599-q5d8b\" (UID: \"7f90213c-484b-4f9d-b064-363887589530\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q5d8b" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276412 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxm7f\" (UniqueName: \"kubernetes.io/projected/58f094f8-8c16-4929-ae94-e52241c5b7d1-kube-api-access-qxm7f\") pod \"cluster-samples-operator-665b6dd947-j665r\" (UID: \"58f094f8-8c16-4929-ae94-e52241c5b7d1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j665r" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276428 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/994432aa-8794-4818-94bd-ce28d97c7163-audit-dir\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276440 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276454 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hcm8\" (UniqueName: \"kubernetes.io/projected/534e137b-1c8d-45fb-9c1b-6804913d309b-kube-api-access-8hcm8\") pod \"controller-manager-879f6c89f-r5ts6\" (UID: \"534e137b-1c8d-45fb-9c1b-6804913d309b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r5ts6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276468 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a883895-9775-4c18-962e-be4df71043c8-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-q79mk\" (UID: \"3a883895-9775-4c18-962e-be4df71043c8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q79mk" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276480 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276496 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276510 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276522 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/534e137b-1c8d-45fb-9c1b-6804913d309b-config\") pod \"controller-manager-879f6c89f-r5ts6\" (UID: \"534e137b-1c8d-45fb-9c1b-6804913d309b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r5ts6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276537 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e814799e-e374-4ad8-96a2-46e4d5defc2f-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-ntvtt\" (UID: \"e814799e-e374-4ad8-96a2-46e4d5defc2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276550 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/18f99d77-1c2d-412b-93ae-1d1fc52f24ab-images\") pod \"machine-api-operator-5694c8668f-ztnsp\" (UID: \"18f99d77-1c2d-412b-93ae-1d1fc52f24ab\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ztnsp" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276563 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41ae564b-da35-45eb-9a27-fbf23bacaf8f-config\") pod \"console-operator-58897d9998-rqftc\" (UID: \"41ae564b-da35-45eb-9a27-fbf23bacaf8f\") " pod="openshift-console-operator/console-operator-58897d9998-rqftc" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276576 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41ae564b-da35-45eb-9a27-fbf23bacaf8f-serving-cert\") pod \"console-operator-58897d9998-rqftc\" (UID: \"41ae564b-da35-45eb-9a27-fbf23bacaf8f\") " pod="openshift-console-operator/console-operator-58897d9998-rqftc" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276587 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/9e9514f9-f14f-4409-8fef-b89a37de70a6-etcd-service-ca\") pod \"etcd-operator-b45778765-9m2sk\" (UID: \"9e9514f9-f14f-4409-8fef-b89a37de70a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9m2sk" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276600 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk4b8\" (UniqueName: \"kubernetes.io/projected/410014cb-6ba0-4610-acc5-38f3db8fcae2-kube-api-access-kk4b8\") pod \"machine-approver-56656f9798-l7z49\" (UID: \"410014cb-6ba0-4610-acc5-38f3db8fcae2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7z49" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.276612 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zph27\" (UniqueName: \"kubernetes.io/projected/4316762b-b889-41c0-aa31-f874bc0fa3fe-kube-api-access-zph27\") pod \"downloads-7954f5f757-p9rps\" (UID: \"4316762b-b889-41c0-aa31-f874bc0fa3fe\") " pod="openshift-console/downloads-7954f5f757-p9rps" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.285362 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e814799e-e374-4ad8-96a2-46e4d5defc2f-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-ntvtt\" (UID: \"e814799e-e374-4ad8-96a2-46e4d5defc2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.285891 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-lktfh"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.288812 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.288854 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18f99d77-1c2d-412b-93ae-1d1fc52f24ab-config\") pod \"machine-api-operator-5694c8668f-ztnsp\" (UID: \"18f99d77-1c2d-412b-93ae-1d1fc52f24ab\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ztnsp" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.293074 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/994432aa-8794-4818-94bd-ce28d97c7163-node-pullsecrets\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.296232 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.296294 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08b0652b-f30f-4c2c-bd1a-75b80a66b144-serving-cert\") pod \"route-controller-manager-6576b87f9c-hzh5b\" (UID: \"08b0652b-f30f-4c2c-bd1a-75b80a66b144\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.299693 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-ltkhr"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.299719 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.300019 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vbcd9"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.300294 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-9775t"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.300622 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lktfh" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.300783 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9775t" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.300795 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-ltkhr" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.300780 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.300927 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vbcd9" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.300932 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.296317 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h642b\" (UniqueName: \"kubernetes.io/projected/08b0652b-f30f-4c2c-bd1a-75b80a66b144-kube-api-access-h642b\") pod \"route-controller-manager-6576b87f9c-hzh5b\" (UID: \"08b0652b-f30f-4c2c-bd1a-75b80a66b144\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.301015 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e814799e-e374-4ad8-96a2-46e4d5defc2f-encryption-config\") pod \"apiserver-7bbb656c7d-ntvtt\" (UID: \"e814799e-e374-4ad8-96a2-46e4d5defc2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.301032 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e7b5892a-8920-4cef-94a1-3736328e9a39-audit-dir\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.301043 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tgbrt"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.301048 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8j9h\" (UniqueName: \"kubernetes.io/projected/16e9fcf3-adc7-4af9-a625-125ff496cfcf-kube-api-access-t8j9h\") pod \"dns-operator-744455d44c-dlj2h\" (UID: \"16e9fcf3-adc7-4af9-a625-125ff496cfcf\") " pod="openshift-dns-operator/dns-operator-744455d44c-dlj2h" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.301327 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6796a9bd-a428-4b0a-8723-d8a6a252de84-console-config\") pod \"console-f9d7485db-xvd4d\" (UID: \"6796a9bd-a428-4b0a-8723-d8a6a252de84\") " pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.301404 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.301402 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6796a9bd-a428-4b0a-8723-d8a6a252de84-oauth-serving-cert\") pod \"console-f9d7485db-xvd4d\" (UID: \"6796a9bd-a428-4b0a-8723-d8a6a252de84\") " pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.301497 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.302015 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tgbrt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.302144 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.302220 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398005-nnldq"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.302711 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-nnldq" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.303749 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2j2z7"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.304297 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2j2z7" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.305083 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.305705 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.306499 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8d5rf"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.307693 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mkb95"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.308076 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mkb95" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.308084 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-5kfl6"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.308239 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8d5rf" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.309535 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-jgl4n"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.309994 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-jgl4n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.313365 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.314007 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.315798 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-kg77n"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.328538 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r5ts6"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.329707 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j665r"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.330705 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-lktfh"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.330988 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q79mk"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.331838 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zr8cm"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.333263 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-t2rfc"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.333540 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.333854 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-rqftc"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.334603 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ggt5f"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.335401 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-q5d8b"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.336320 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.337147 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-xvd4d"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.337991 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-78zq2"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.338712 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-9m2sk"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.339509 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8xfrw"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.340317 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-p9rps"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.342215 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vbcd9"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.344771 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tgbrt"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.347090 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-dnz8k"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.348652 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-dlj2h"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.349696 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-ltkhr"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.350503 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qt4fl"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.351355 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ccrq2"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.352158 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-8wt84"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.352981 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xwb9l"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.353070 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.353848 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-fjv2r"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.354688 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r779q"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.355563 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8llfp"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.356332 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398005-nnldq"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.357165 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-9775t"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.357934 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2p9ch"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.358742 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-mfl2b"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.358919 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-2p9ch" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.359620 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2j2z7"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.359717 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-mfl2b" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.360362 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8d5rf"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.361182 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mkb95"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.362006 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-mfl2b"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.362821 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2p9ch"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.373116 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.393481 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.402576 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/41ae564b-da35-45eb-9a27-fbf23bacaf8f-trusted-ca\") pod \"console-operator-58897d9998-rqftc\" (UID: \"41ae564b-da35-45eb-9a27-fbf23bacaf8f\") " pod="openshift-console-operator/console-operator-58897d9998-rqftc" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.402605 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e9514f9-f14f-4409-8fef-b89a37de70a6-serving-cert\") pod \"etcd-operator-b45778765-9m2sk\" (UID: \"9e9514f9-f14f-4409-8fef-b89a37de70a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9m2sk" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.402624 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8xcm\" (UniqueName: \"kubernetes.io/projected/9e9514f9-f14f-4409-8fef-b89a37de70a6-kube-api-access-f8xcm\") pod \"etcd-operator-b45778765-9m2sk\" (UID: \"9e9514f9-f14f-4409-8fef-b89a37de70a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9m2sk" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.402653 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/994432aa-8794-4818-94bd-ce28d97c7163-encryption-config\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.402672 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e814799e-e374-4ad8-96a2-46e4d5defc2f-serving-cert\") pod \"apiserver-7bbb656c7d-ntvtt\" (UID: \"e814799e-e374-4ad8-96a2-46e4d5defc2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.402695 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6mxz\" (UniqueName: \"kubernetes.io/projected/41ae564b-da35-45eb-9a27-fbf23bacaf8f-kube-api-access-b6mxz\") pod \"console-operator-58897d9998-rqftc\" (UID: \"41ae564b-da35-45eb-9a27-fbf23bacaf8f\") " pod="openshift-console-operator/console-operator-58897d9998-rqftc" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.402710 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/9e9514f9-f14f-4409-8fef-b89a37de70a6-etcd-ca\") pod \"etcd-operator-b45778765-9m2sk\" (UID: \"9e9514f9-f14f-4409-8fef-b89a37de70a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9m2sk" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.402730 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znnk8\" (UniqueName: \"kubernetes.io/projected/266931fd-4205-4e31-9ded-9882a2a41921-kube-api-access-znnk8\") pod \"migrator-59844c95c7-9775t\" (UID: \"266931fd-4205-4e31-9ded-9882a2a41921\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9775t" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.402751 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b3e2ea56-3525-4e78-a7c8-289a9a2406e6-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-8llfp\" (UID: \"b3e2ea56-3525-4e78-a7c8-289a9a2406e6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8llfp" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.402768 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.402785 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.402800 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/16e9fcf3-adc7-4af9-a625-125ff496cfcf-metrics-tls\") pod \"dns-operator-744455d44c-dlj2h\" (UID: \"16e9fcf3-adc7-4af9-a625-125ff496cfcf\") " pod="openshift-dns-operator/dns-operator-744455d44c-dlj2h" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.402815 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/534e137b-1c8d-45fb-9c1b-6804913d309b-client-ca\") pod \"controller-manager-879f6c89f-r5ts6\" (UID: \"534e137b-1c8d-45fb-9c1b-6804913d309b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r5ts6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.402851 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/410014cb-6ba0-4610-acc5-38f3db8fcae2-machine-approver-tls\") pod \"machine-approver-56656f9798-l7z49\" (UID: \"410014cb-6ba0-4610-acc5-38f3db8fcae2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7z49" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.402882 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e814799e-e374-4ad8-96a2-46e4d5defc2f-audit-policies\") pod \"apiserver-7bbb656c7d-ntvtt\" (UID: \"e814799e-e374-4ad8-96a2-46e4d5defc2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.402902 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.402919 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8hn2\" (UniqueName: \"kubernetes.io/projected/7f90213c-484b-4f9d-b064-363887589530-kube-api-access-g8hn2\") pod \"authentication-operator-69f744f599-q5d8b\" (UID: \"7f90213c-484b-4f9d-b064-363887589530\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q5d8b" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.402935 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxm7f\" (UniqueName: \"kubernetes.io/projected/58f094f8-8c16-4929-ae94-e52241c5b7d1-kube-api-access-qxm7f\") pod \"cluster-samples-operator-665b6dd947-j665r\" (UID: \"58f094f8-8c16-4929-ae94-e52241c5b7d1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j665r" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.402950 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/994432aa-8794-4818-94bd-ce28d97c7163-audit-dir\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.402964 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.402978 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hcm8\" (UniqueName: \"kubernetes.io/projected/534e137b-1c8d-45fb-9c1b-6804913d309b-kube-api-access-8hcm8\") pod \"controller-manager-879f6c89f-r5ts6\" (UID: \"534e137b-1c8d-45fb-9c1b-6804913d309b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r5ts6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.402995 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2879a16a-7ab6-413a-a8a3-ecfa7ffd4ac7-metrics-tls\") pod \"ingress-operator-5b745b69d9-dnz8k\" (UID: \"2879a16a-7ab6-413a-a8a3-ecfa7ffd4ac7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dnz8k" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403010 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb-srv-cert\") pod \"catalog-operator-68c6474976-mkb95\" (UID: \"0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mkb95" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403028 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403042 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/534e137b-1c8d-45fb-9c1b-6804913d309b-config\") pod \"controller-manager-879f6c89f-r5ts6\" (UID: \"534e137b-1c8d-45fb-9c1b-6804913d309b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r5ts6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403057 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e814799e-e374-4ad8-96a2-46e4d5defc2f-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-ntvtt\" (UID: \"e814799e-e374-4ad8-96a2-46e4d5defc2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403070 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/18f99d77-1c2d-412b-93ae-1d1fc52f24ab-images\") pod \"machine-api-operator-5694c8668f-ztnsp\" (UID: \"18f99d77-1c2d-412b-93ae-1d1fc52f24ab\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ztnsp" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403089 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9v4zr\" (UniqueName: \"kubernetes.io/projected/04abe9e5-6ffe-429d-9274-73d133fd0d52-kube-api-access-9v4zr\") pod \"package-server-manager-789f6589d5-r779q\" (UID: \"04abe9e5-6ffe-429d-9274-73d133fd0d52\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r779q" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403105 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41ae564b-da35-45eb-9a27-fbf23bacaf8f-serving-cert\") pod \"console-operator-58897d9998-rqftc\" (UID: \"41ae564b-da35-45eb-9a27-fbf23bacaf8f\") " pod="openshift-console-operator/console-operator-58897d9998-rqftc" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403120 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18f99d77-1c2d-412b-93ae-1d1fc52f24ab-config\") pod \"machine-api-operator-5694c8668f-ztnsp\" (UID: \"18f99d77-1c2d-412b-93ae-1d1fc52f24ab\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ztnsp" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403136 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbvd4\" (UniqueName: \"kubernetes.io/projected/a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba-kube-api-access-hbvd4\") pod \"router-default-5444994796-9gwbh\" (UID: \"a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba\") " pod="openshift-ingress/router-default-5444994796-9gwbh" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403150 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk4b8\" (UniqueName: \"kubernetes.io/projected/410014cb-6ba0-4610-acc5-38f3db8fcae2-kube-api-access-kk4b8\") pod \"machine-approver-56656f9798-l7z49\" (UID: \"410014cb-6ba0-4610-acc5-38f3db8fcae2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7z49" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403166 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb-profile-collector-cert\") pod \"catalog-operator-68c6474976-mkb95\" (UID: \"0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mkb95" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403182 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403199 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e814799e-e374-4ad8-96a2-46e4d5defc2f-encryption-config\") pod \"apiserver-7bbb656c7d-ntvtt\" (UID: \"e814799e-e374-4ad8-96a2-46e4d5defc2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403213 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6796a9bd-a428-4b0a-8723-d8a6a252de84-console-config\") pod \"console-f9d7485db-xvd4d\" (UID: \"6796a9bd-a428-4b0a-8723-d8a6a252de84\") " pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403228 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6796a9bd-a428-4b0a-8723-d8a6a252de84-oauth-serving-cert\") pod \"console-f9d7485db-xvd4d\" (UID: \"6796a9bd-a428-4b0a-8723-d8a6a252de84\") " pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403242 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n658h\" (UniqueName: \"kubernetes.io/projected/3d1342fa-3bb4-4924-a36b-e370aec66cd3-kube-api-access-n658h\") pod \"machine-config-server-jgl4n\" (UID: \"3d1342fa-3bb4-4924-a36b-e370aec66cd3\") " pod="openshift-machine-config-operator/machine-config-server-jgl4n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403257 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8j9h\" (UniqueName: \"kubernetes.io/projected/16e9fcf3-adc7-4af9-a625-125ff496cfcf-kube-api-access-t8j9h\") pod \"dns-operator-744455d44c-dlj2h\" (UID: \"16e9fcf3-adc7-4af9-a625-125ff496cfcf\") " pod="openshift-dns-operator/dns-operator-744455d44c-dlj2h" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403273 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k45sh\" (UniqueName: \"kubernetes.io/projected/09d28a3a-5b25-4e31-90d6-17d9319cd11c-kube-api-access-k45sh\") pod \"machine-config-controller-84d6567774-t2rfc\" (UID: \"09d28a3a-5b25-4e31-90d6-17d9319cd11c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t2rfc" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403289 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2879a16a-7ab6-413a-a8a3-ecfa7ffd4ac7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-dnz8k\" (UID: \"2879a16a-7ab6-413a-a8a3-ecfa7ffd4ac7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dnz8k" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403305 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/534e137b-1c8d-45fb-9c1b-6804913d309b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-r5ts6\" (UID: \"534e137b-1c8d-45fb-9c1b-6804913d309b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r5ts6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403322 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnflg\" (UniqueName: \"kubernetes.io/projected/994432aa-8794-4818-94bd-ce28d97c7163-kube-api-access-qnflg\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403337 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b7acf5d-cae8-429a-8661-ec3abcf056c0-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-xwb9l\" (UID: \"1b7acf5d-cae8-429a-8661-ec3abcf056c0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xwb9l" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403351 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3a922bdd-f7be-41d2-8626-9352ec57d9ce-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zr8cm\" (UID: \"3a922bdd-f7be-41d2-8626-9352ec57d9ce\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zr8cm" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403366 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d1d0d430-031f-448e-9763-69deb17abc4a-config-volume\") pod \"collect-profiles-29398005-nnldq\" (UID: \"d1d0d430-031f-448e-9763-69deb17abc4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-nnldq" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403380 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/994432aa-8794-4818-94bd-ce28d97c7163-etcd-client\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403394 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/410014cb-6ba0-4610-acc5-38f3db8fcae2-config\") pod \"machine-approver-56656f9798-l7z49\" (UID: \"410014cb-6ba0-4610-acc5-38f3db8fcae2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7z49" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403409 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4zsf\" (UniqueName: \"kubernetes.io/projected/ae746e9c-3187-4a3b-a439-b5ef25b37caf-kube-api-access-x4zsf\") pod \"control-plane-machine-set-operator-78cbb6b69f-2j2z7\" (UID: \"ae746e9c-3187-4a3b-a439-b5ef25b37caf\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2j2z7" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403426 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmdwc\" (UniqueName: \"kubernetes.io/projected/80d890b8-c797-4392-a48a-7894059147d4-kube-api-access-bmdwc\") pod \"openshift-controller-manager-operator-756b6f6bc6-qt4fl\" (UID: \"80d890b8-c797-4392-a48a-7894059147d4\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qt4fl" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403433 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/534e137b-1c8d-45fb-9c1b-6804913d309b-client-ca\") pod \"controller-manager-879f6c89f-r5ts6\" (UID: \"534e137b-1c8d-45fb-9c1b-6804913d309b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r5ts6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403442 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/994432aa-8794-4818-94bd-ce28d97c7163-serving-cert\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403457 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6796a9bd-a428-4b0a-8723-d8a6a252de84-console-serving-cert\") pod \"console-f9d7485db-xvd4d\" (UID: \"6796a9bd-a428-4b0a-8723-d8a6a252de84\") " pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403473 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/09d28a3a-5b25-4e31-90d6-17d9319cd11c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-t2rfc\" (UID: \"09d28a3a-5b25-4e31-90d6-17d9319cd11c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t2rfc" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403489 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/04abe9e5-6ffe-429d-9274-73d133fd0d52-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-r779q\" (UID: \"04abe9e5-6ffe-429d-9274-73d133fd0d52\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r779q" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403503 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a130a936-3db2-4455-bb63-fcea37eda82c-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-8xfrw\" (UID: \"a130a936-3db2-4455-bb63-fcea37eda82c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8xfrw" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403518 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e9514f9-f14f-4409-8fef-b89a37de70a6-config\") pod \"etcd-operator-b45778765-9m2sk\" (UID: \"9e9514f9-f14f-4409-8fef-b89a37de70a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9m2sk" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403534 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e814799e-e374-4ad8-96a2-46e4d5defc2f-etcd-client\") pod \"apiserver-7bbb656c7d-ntvtt\" (UID: \"e814799e-e374-4ad8-96a2-46e4d5defc2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403549 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a922bdd-f7be-41d2-8626-9352ec57d9ce-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zr8cm\" (UID: \"3a922bdd-f7be-41d2-8626-9352ec57d9ce\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zr8cm" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403565 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b3e2ea56-3525-4e78-a7c8-289a9a2406e6-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-8llfp\" (UID: \"b3e2ea56-3525-4e78-a7c8-289a9a2406e6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8llfp" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403579 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqq59\" (UniqueName: \"kubernetes.io/projected/337f295d-6c74-4850-bb14-125192de4385-kube-api-access-pqq59\") pod \"marketplace-operator-79b997595-tgbrt\" (UID: \"337f295d-6c74-4850-bb14-125192de4385\") " pod="openshift-marketplace/marketplace-operator-79b997595-tgbrt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403596 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80d890b8-c797-4392-a48a-7894059147d4-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-qt4fl\" (UID: \"80d890b8-c797-4392-a48a-7894059147d4\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qt4fl" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403611 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn86q\" (UniqueName: \"kubernetes.io/projected/833dd52a-3337-431a-af83-d75240e6ab5c-kube-api-access-fn86q\") pod \"packageserver-d55dfcdfc-78zq2\" (UID: \"833dd52a-3337-431a-af83-d75240e6ab5c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-78zq2" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403629 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/58f094f8-8c16-4929-ae94-e52241c5b7d1-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-j665r\" (UID: \"58f094f8-8c16-4929-ae94-e52241c5b7d1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j665r" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403659 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/50e7a859-2c04-4eff-889d-633ced116f34-auth-proxy-config\") pod \"machine-config-operator-74547568cd-lktfh\" (UID: \"50e7a859-2c04-4eff-889d-633ced116f34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lktfh" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403676 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403691 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f90213c-484b-4f9d-b064-363887589530-serving-cert\") pod \"authentication-operator-69f744f599-q5d8b\" (UID: \"7f90213c-484b-4f9d-b064-363887589530\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q5d8b" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403706 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/833dd52a-3337-431a-af83-d75240e6ab5c-tmpfs\") pod \"packageserver-d55dfcdfc-78zq2\" (UID: \"833dd52a-3337-431a-af83-d75240e6ab5c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-78zq2" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403720 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6796a9bd-a428-4b0a-8723-d8a6a252de84-service-ca\") pod \"console-f9d7485db-xvd4d\" (UID: \"6796a9bd-a428-4b0a-8723-d8a6a252de84\") " pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403737 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35b44360-64aa-45b1-9141-15d5f69b384a-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ggt5f\" (UID: \"35b44360-64aa-45b1-9141-15d5f69b384a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ggt5f" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403409 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/41ae564b-da35-45eb-9a27-fbf23bacaf8f-trusted-ca\") pod \"console-operator-58897d9998-rqftc\" (UID: \"41ae564b-da35-45eb-9a27-fbf23bacaf8f\") " pod="openshift-console-operator/console-operator-58897d9998-rqftc" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403753 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m24rv\" (UniqueName: \"kubernetes.io/projected/528f6e87-f9e3-46f4-b40b-127b4092b8aa-kube-api-access-m24rv\") pod \"service-ca-9c57cc56f-ltkhr\" (UID: \"528f6e87-f9e3-46f4-b40b-127b4092b8aa\") " pod="openshift-service-ca/service-ca-9c57cc56f-ltkhr" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403831 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z57mp\" (UniqueName: \"kubernetes.io/projected/e7b5892a-8920-4cef-94a1-3736328e9a39-kube-api-access-z57mp\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403854 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l6xn\" (UniqueName: \"kubernetes.io/projected/e814799e-e374-4ad8-96a2-46e4d5defc2f-kube-api-access-6l6xn\") pod \"apiserver-7bbb656c7d-ntvtt\" (UID: \"e814799e-e374-4ad8-96a2-46e4d5defc2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403889 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/994432aa-8794-4818-94bd-ce28d97c7163-config\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403910 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/994432aa-8794-4818-94bd-ce28d97c7163-image-import-ca\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403929 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403948 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9e9514f9-f14f-4409-8fef-b89a37de70a6-etcd-client\") pod \"etcd-operator-b45778765-9m2sk\" (UID: \"9e9514f9-f14f-4409-8fef-b89a37de70a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9m2sk" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403967 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6796a9bd-a428-4b0a-8723-d8a6a252de84-trusted-ca-bundle\") pod \"console-f9d7485db-xvd4d\" (UID: \"6796a9bd-a428-4b0a-8723-d8a6a252de84\") " pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.403984 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e814799e-e374-4ad8-96a2-46e4d5defc2f-audit-dir\") pod \"apiserver-7bbb656c7d-ntvtt\" (UID: \"e814799e-e374-4ad8-96a2-46e4d5defc2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404006 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/50e7a859-2c04-4eff-889d-633ced116f34-images\") pod \"machine-config-operator-74547568cd-lktfh\" (UID: \"50e7a859-2c04-4eff-889d-633ced116f34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lktfh" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404018 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18f99d77-1c2d-412b-93ae-1d1fc52f24ab-config\") pod \"machine-api-operator-5694c8668f-ztnsp\" (UID: \"18f99d77-1c2d-412b-93ae-1d1fc52f24ab\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ztnsp" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404025 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/337f295d-6c74-4850-bb14-125192de4385-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tgbrt\" (UID: \"337f295d-6c74-4850-bb14-125192de4385\") " pod="openshift-marketplace/marketplace-operator-79b997595-tgbrt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404048 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2879a16a-7ab6-413a-a8a3-ecfa7ffd4ac7-trusted-ca\") pod \"ingress-operator-5b745b69d9-dnz8k\" (UID: \"2879a16a-7ab6-413a-a8a3-ecfa7ffd4ac7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dnz8k" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404081 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndlbq\" (UniqueName: \"kubernetes.io/projected/e417e794-f6b2-4607-b2ce-3a737e8770b4-kube-api-access-ndlbq\") pod \"openshift-config-operator-7777fb866f-fjv2r\" (UID: \"e417e794-f6b2-4607-b2ce-3a737e8770b4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fjv2r" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404103 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f90213c-484b-4f9d-b064-363887589530-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-q5d8b\" (UID: \"7f90213c-484b-4f9d-b064-363887589530\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q5d8b" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404123 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/35b44360-64aa-45b1-9141-15d5f69b384a-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ggt5f\" (UID: \"35b44360-64aa-45b1-9141-15d5f69b384a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ggt5f" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404149 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/534e137b-1c8d-45fb-9c1b-6804913d309b-serving-cert\") pod \"controller-manager-879f6c89f-r5ts6\" (UID: \"534e137b-1c8d-45fb-9c1b-6804913d309b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r5ts6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404168 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e417e794-f6b2-4607-b2ce-3a737e8770b4-serving-cert\") pod \"openshift-config-operator-7777fb866f-fjv2r\" (UID: \"e417e794-f6b2-4607-b2ce-3a737e8770b4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fjv2r" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404183 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfjqs\" (UniqueName: \"kubernetes.io/projected/3a883895-9775-4c18-962e-be4df71043c8-kube-api-access-sfjqs\") pod \"openshift-apiserver-operator-796bbdcf4f-q79mk\" (UID: \"3a883895-9775-4c18-962e-be4df71043c8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q79mk" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404198 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba-default-certificate\") pod \"router-default-5444994796-9gwbh\" (UID: \"a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba\") " pod="openshift-ingress/router-default-5444994796-9gwbh" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404261 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a883895-9775-4c18-962e-be4df71043c8-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-q79mk\" (UID: \"3a883895-9775-4c18-962e-be4df71043c8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q79mk" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404280 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80d890b8-c797-4392-a48a-7894059147d4-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-qt4fl\" (UID: \"80d890b8-c797-4392-a48a-7894059147d4\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qt4fl" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404297 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404313 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404331 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2zdw\" (UniqueName: \"kubernetes.io/projected/a91fbd9b-8264-4ac4-b1bd-d62143b7a4d2-kube-api-access-f2zdw\") pod \"multus-admission-controller-857f4d67dd-8wt84\" (UID: \"a91fbd9b-8264-4ac4-b1bd-d62143b7a4d2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8wt84" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404347 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41ae564b-da35-45eb-9a27-fbf23bacaf8f-config\") pod \"console-operator-58897d9998-rqftc\" (UID: \"41ae564b-da35-45eb-9a27-fbf23bacaf8f\") " pod="openshift-console-operator/console-operator-58897d9998-rqftc" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404363 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/9e9514f9-f14f-4409-8fef-b89a37de70a6-etcd-service-ca\") pod \"etcd-operator-b45778765-9m2sk\" (UID: \"9e9514f9-f14f-4409-8fef-b89a37de70a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9m2sk" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404378 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3d1342fa-3bb4-4924-a36b-e370aec66cd3-certs\") pod \"machine-config-server-jgl4n\" (UID: \"3d1342fa-3bb4-4924-a36b-e370aec66cd3\") " pod="openshift-machine-config-operator/machine-config-server-jgl4n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404394 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szjqc\" (UniqueName: \"kubernetes.io/projected/0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb-kube-api-access-szjqc\") pod \"catalog-operator-68c6474976-mkb95\" (UID: \"0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mkb95" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404410 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zph27\" (UniqueName: \"kubernetes.io/projected/4316762b-b889-41c0-aa31-f874bc0fa3fe-kube-api-access-zph27\") pod \"downloads-7954f5f757-p9rps\" (UID: \"4316762b-b889-41c0-aa31-f874bc0fa3fe\") " pod="openshift-console/downloads-7954f5f757-p9rps" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404428 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e814799e-e374-4ad8-96a2-46e4d5defc2f-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-ntvtt\" (UID: \"e814799e-e374-4ad8-96a2-46e4d5defc2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404442 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d1d0d430-031f-448e-9763-69deb17abc4a-secret-volume\") pod \"collect-profiles-29398005-nnldq\" (UID: \"d1d0d430-031f-448e-9763-69deb17abc4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-nnldq" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404458 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/994432aa-8794-4818-94bd-ce28d97c7163-node-pullsecrets\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404491 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h642b\" (UniqueName: \"kubernetes.io/projected/08b0652b-f30f-4c2c-bd1a-75b80a66b144-kube-api-access-h642b\") pod \"route-controller-manager-6576b87f9c-hzh5b\" (UID: \"08b0652b-f30f-4c2c-bd1a-75b80a66b144\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404506 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/528f6e87-f9e3-46f4-b40b-127b4092b8aa-signing-cabundle\") pod \"service-ca-9c57cc56f-ltkhr\" (UID: \"528f6e87-f9e3-46f4-b40b-127b4092b8aa\") " pod="openshift-service-ca/service-ca-9c57cc56f-ltkhr" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404521 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08b0652b-f30f-4c2c-bd1a-75b80a66b144-serving-cert\") pod \"route-controller-manager-6576b87f9c-hzh5b\" (UID: \"08b0652b-f30f-4c2c-bd1a-75b80a66b144\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404537 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e7b5892a-8920-4cef-94a1-3736328e9a39-audit-dir\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404553 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwb8n\" (UniqueName: \"kubernetes.io/projected/b3e2ea56-3525-4e78-a7c8-289a9a2406e6-kube-api-access-bwb8n\") pod \"cluster-image-registry-operator-dc59b4c8b-8llfp\" (UID: \"b3e2ea56-3525-4e78-a7c8-289a9a2406e6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8llfp" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404567 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/994432aa-8794-4818-94bd-ce28d97c7163-audit\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404582 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/09d28a3a-5b25-4e31-90d6-17d9319cd11c-proxy-tls\") pod \"machine-config-controller-84d6567774-t2rfc\" (UID: \"09d28a3a-5b25-4e31-90d6-17d9319cd11c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t2rfc" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404597 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9cjx\" (UniqueName: \"kubernetes.io/projected/35b44360-64aa-45b1-9141-15d5f69b384a-kube-api-access-z9cjx\") pod \"kube-storage-version-migrator-operator-b67b599dd-ggt5f\" (UID: \"35b44360-64aa-45b1-9141-15d5f69b384a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ggt5f" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404612 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/337f295d-6c74-4850-bb14-125192de4385-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tgbrt\" (UID: \"337f295d-6c74-4850-bb14-125192de4385\") " pod="openshift-marketplace/marketplace-operator-79b997595-tgbrt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404629 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba-service-ca-bundle\") pod \"router-default-5444994796-9gwbh\" (UID: \"a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba\") " pod="openshift-ingress/router-default-5444994796-9gwbh" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404632 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404660 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a130a936-3db2-4455-bb63-fcea37eda82c-config\") pod \"kube-controller-manager-operator-78b949d7b-8xfrw\" (UID: \"a130a936-3db2-4455-bb63-fcea37eda82c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8xfrw" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404679 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/e417e794-f6b2-4607-b2ce-3a737e8770b4-available-featuregates\") pod \"openshift-config-operator-7777fb866f-fjv2r\" (UID: \"e417e794-f6b2-4607-b2ce-3a737e8770b4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fjv2r" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404706 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/410014cb-6ba0-4610-acc5-38f3db8fcae2-auth-proxy-config\") pod \"machine-approver-56656f9798-l7z49\" (UID: \"410014cb-6ba0-4610-acc5-38f3db8fcae2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7z49" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404755 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a883895-9775-4c18-962e-be4df71043c8-config\") pod \"openshift-apiserver-operator-796bbdcf4f-q79mk\" (UID: \"3a883895-9775-4c18-962e-be4df71043c8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q79mk" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404777 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/ae746e9c-3187-4a3b-a439-b5ef25b37caf-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-2j2z7\" (UID: \"ae746e9c-3187-4a3b-a439-b5ef25b37caf\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2j2z7" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404796 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/833dd52a-3337-431a-af83-d75240e6ab5c-apiservice-cert\") pod \"packageserver-d55dfcdfc-78zq2\" (UID: \"833dd52a-3337-431a-af83-d75240e6ab5c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-78zq2" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404811 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba-metrics-certs\") pod \"router-default-5444994796-9gwbh\" (UID: \"a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba\") " pod="openshift-ingress/router-default-5444994796-9gwbh" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404826 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3d1342fa-3bb4-4924-a36b-e370aec66cd3-node-bootstrap-token\") pod \"machine-config-server-jgl4n\" (UID: \"3d1342fa-3bb4-4924-a36b-e370aec66cd3\") " pod="openshift-machine-config-operator/machine-config-server-jgl4n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404841 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba-stats-auth\") pod \"router-default-5444994796-9gwbh\" (UID: \"a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba\") " pod="openshift-ingress/router-default-5444994796-9gwbh" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404858 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/994432aa-8794-4818-94bd-ce28d97c7163-etcd-serving-ca\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404886 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6796a9bd-a428-4b0a-8723-d8a6a252de84-console-oauth-config\") pod \"console-f9d7485db-xvd4d\" (UID: \"6796a9bd-a428-4b0a-8723-d8a6a252de84\") " pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404903 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5hv8\" (UniqueName: \"kubernetes.io/projected/50e7a859-2c04-4eff-889d-633ced116f34-kube-api-access-l5hv8\") pod \"machine-config-operator-74547568cd-lktfh\" (UID: \"50e7a859-2c04-4eff-889d-633ced116f34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lktfh" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404919 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzshk\" (UniqueName: \"kubernetes.io/projected/d1d0d430-031f-448e-9763-69deb17abc4a-kube-api-access-nzshk\") pod \"collect-profiles-29398005-nnldq\" (UID: \"d1d0d430-031f-448e-9763-69deb17abc4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-nnldq" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404936 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a130a936-3db2-4455-bb63-fcea37eda82c-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-8xfrw\" (UID: \"a130a936-3db2-4455-bb63-fcea37eda82c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8xfrw" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404955 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.404971 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b3e2ea56-3525-4e78-a7c8-289a9a2406e6-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-8llfp\" (UID: \"b3e2ea56-3525-4e78-a7c8-289a9a2406e6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8llfp" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.405005 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08b0652b-f30f-4c2c-bd1a-75b80a66b144-config\") pod \"route-controller-manager-6576b87f9c-hzh5b\" (UID: \"08b0652b-f30f-4c2c-bd1a-75b80a66b144\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.405020 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f90213c-484b-4f9d-b064-363887589530-config\") pod \"authentication-operator-69f744f599-q5d8b\" (UID: \"7f90213c-484b-4f9d-b064-363887589530\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q5d8b" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.405037 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/833dd52a-3337-431a-af83-d75240e6ab5c-webhook-cert\") pod \"packageserver-d55dfcdfc-78zq2\" (UID: \"833dd52a-3337-431a-af83-d75240e6ab5c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-78zq2" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.405054 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/e417e794-f6b2-4607-b2ce-3a737e8770b4-available-featuregates\") pod \"openshift-config-operator-7777fb866f-fjv2r\" (UID: \"e417e794-f6b2-4607-b2ce-3a737e8770b4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fjv2r" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.405070 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a922bdd-f7be-41d2-8626-9352ec57d9ce-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zr8cm\" (UID: \"3a922bdd-f7be-41d2-8626-9352ec57d9ce\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zr8cm" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.405101 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a91fbd9b-8264-4ac4-b1bd-d62143b7a4d2-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-8wt84\" (UID: \"a91fbd9b-8264-4ac4-b1bd-d62143b7a4d2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8wt84" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.405605 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/410014cb-6ba0-4610-acc5-38f3db8fcae2-config\") pod \"machine-approver-56656f9798-l7z49\" (UID: \"410014cb-6ba0-4610-acc5-38f3db8fcae2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7z49" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.405718 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5658\" (UniqueName: \"kubernetes.io/projected/2879a16a-7ab6-413a-a8a3-ecfa7ffd4ac7-kube-api-access-w5658\") pod \"ingress-operator-5b745b69d9-dnz8k\" (UID: \"2879a16a-7ab6-413a-a8a3-ecfa7ffd4ac7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dnz8k" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.405743 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/08b0652b-f30f-4c2c-bd1a-75b80a66b144-client-ca\") pod \"route-controller-manager-6576b87f9c-hzh5b\" (UID: \"08b0652b-f30f-4c2c-bd1a-75b80a66b144\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.405787 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f90213c-484b-4f9d-b064-363887589530-service-ca-bundle\") pod \"authentication-operator-69f744f599-q5d8b\" (UID: \"7f90213c-484b-4f9d-b064-363887589530\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q5d8b" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.405815 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e7b5892a-8920-4cef-94a1-3736328e9a39-audit-policies\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.405855 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b7acf5d-cae8-429a-8661-ec3abcf056c0-config\") pod \"kube-apiserver-operator-766d6c64bb-xwb9l\" (UID: \"1b7acf5d-cae8-429a-8661-ec3abcf056c0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xwb9l" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.405884 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdt5m\" (UniqueName: \"kubernetes.io/projected/18f99d77-1c2d-412b-93ae-1d1fc52f24ab-kube-api-access-vdt5m\") pod \"machine-api-operator-5694c8668f-ztnsp\" (UID: \"18f99d77-1c2d-412b-93ae-1d1fc52f24ab\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ztnsp" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.405904 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1b7acf5d-cae8-429a-8661-ec3abcf056c0-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-xwb9l\" (UID: \"1b7acf5d-cae8-429a-8661-ec3abcf056c0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xwb9l" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.405939 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/994432aa-8794-4818-94bd-ce28d97c7163-trusted-ca-bundle\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.405956 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fjff\" (UniqueName: \"kubernetes.io/projected/6796a9bd-a428-4b0a-8723-d8a6a252de84-kube-api-access-8fjff\") pod \"console-f9d7485db-xvd4d\" (UID: \"6796a9bd-a428-4b0a-8723-d8a6a252de84\") " pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.405971 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/18f99d77-1c2d-412b-93ae-1d1fc52f24ab-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-ztnsp\" (UID: \"18f99d77-1c2d-412b-93ae-1d1fc52f24ab\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ztnsp" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.405985 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/50e7a859-2c04-4eff-889d-633ced116f34-proxy-tls\") pod \"machine-config-operator-74547568cd-lktfh\" (UID: \"50e7a859-2c04-4eff-889d-633ced116f34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lktfh" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.406020 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/528f6e87-f9e3-46f4-b40b-127b4092b8aa-signing-key\") pod \"service-ca-9c57cc56f-ltkhr\" (UID: \"528f6e87-f9e3-46f4-b40b-127b4092b8aa\") " pod="openshift-service-ca/service-ca-9c57cc56f-ltkhr" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.406164 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/994432aa-8794-4818-94bd-ce28d97c7163-config\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.406493 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/994432aa-8794-4818-94bd-ce28d97c7163-encryption-config\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.406777 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/994432aa-8794-4818-94bd-ce28d97c7163-serving-cert\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.406961 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/994432aa-8794-4818-94bd-ce28d97c7163-image-import-ca\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.407027 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b3e2ea56-3525-4e78-a7c8-289a9a2406e6-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-8llfp\" (UID: \"b3e2ea56-3525-4e78-a7c8-289a9a2406e6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8llfp" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.407255 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/994432aa-8794-4818-94bd-ce28d97c7163-node-pullsecrets\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.407357 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/16e9fcf3-adc7-4af9-a625-125ff496cfcf-metrics-tls\") pod \"dns-operator-744455d44c-dlj2h\" (UID: \"16e9fcf3-adc7-4af9-a625-125ff496cfcf\") " pod="openshift-dns-operator/dns-operator-744455d44c-dlj2h" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.407675 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/410014cb-6ba0-4610-acc5-38f3db8fcae2-auth-proxy-config\") pod \"machine-approver-56656f9798-l7z49\" (UID: \"410014cb-6ba0-4610-acc5-38f3db8fcae2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7z49" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.407805 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.408158 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a883895-9775-4c18-962e-be4df71043c8-config\") pod \"openshift-apiserver-operator-796bbdcf4f-q79mk\" (UID: \"3a883895-9775-4c18-962e-be4df71043c8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q79mk" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.408363 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e814799e-e374-4ad8-96a2-46e4d5defc2f-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-ntvtt\" (UID: \"e814799e-e374-4ad8-96a2-46e4d5defc2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.408760 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6796a9bd-a428-4b0a-8723-d8a6a252de84-console-config\") pod \"console-f9d7485db-xvd4d\" (UID: \"6796a9bd-a428-4b0a-8723-d8a6a252de84\") " pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.408897 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/58f094f8-8c16-4929-ae94-e52241c5b7d1-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-j665r\" (UID: \"58f094f8-8c16-4929-ae94-e52241c5b7d1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j665r" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.409000 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6796a9bd-a428-4b0a-8723-d8a6a252de84-oauth-serving-cert\") pod \"console-f9d7485db-xvd4d\" (UID: \"6796a9bd-a428-4b0a-8723-d8a6a252de84\") " pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.409213 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.409538 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/9e9514f9-f14f-4409-8fef-b89a37de70a6-etcd-ca\") pod \"etcd-operator-b45778765-9m2sk\" (UID: \"9e9514f9-f14f-4409-8fef-b89a37de70a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9m2sk" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.409603 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a883895-9775-4c18-962e-be4df71043c8-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-q79mk\" (UID: \"3a883895-9775-4c18-962e-be4df71043c8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q79mk" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.409876 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9e9514f9-f14f-4409-8fef-b89a37de70a6-etcd-client\") pod \"etcd-operator-b45778765-9m2sk\" (UID: \"9e9514f9-f14f-4409-8fef-b89a37de70a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9m2sk" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.410041 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/994432aa-8794-4818-94bd-ce28d97c7163-etcd-serving-ca\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.410082 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/18f99d77-1c2d-412b-93ae-1d1fc52f24ab-images\") pod \"machine-api-operator-5694c8668f-ztnsp\" (UID: \"18f99d77-1c2d-412b-93ae-1d1fc52f24ab\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ztnsp" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.410129 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.410308 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6796a9bd-a428-4b0a-8723-d8a6a252de84-service-ca\") pod \"console-f9d7485db-xvd4d\" (UID: \"6796a9bd-a428-4b0a-8723-d8a6a252de84\") " pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.410566 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/994432aa-8794-4818-94bd-ce28d97c7163-audit-dir\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.410587 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6796a9bd-a428-4b0a-8723-d8a6a252de84-trusted-ca-bundle\") pod \"console-f9d7485db-xvd4d\" (UID: \"6796a9bd-a428-4b0a-8723-d8a6a252de84\") " pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.410613 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e814799e-e374-4ad8-96a2-46e4d5defc2f-audit-dir\") pod \"apiserver-7bbb656c7d-ntvtt\" (UID: \"e814799e-e374-4ad8-96a2-46e4d5defc2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.410686 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e7b5892a-8920-4cef-94a1-3736328e9a39-audit-dir\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.410897 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/994432aa-8794-4818-94bd-ce28d97c7163-audit\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.411037 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.411416 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f90213c-484b-4f9d-b064-363887589530-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-q5d8b\" (UID: \"7f90213c-484b-4f9d-b064-363887589530\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q5d8b" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.411458 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e814799e-e374-4ad8-96a2-46e4d5defc2f-audit-policies\") pod \"apiserver-7bbb656c7d-ntvtt\" (UID: \"e814799e-e374-4ad8-96a2-46e4d5defc2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.411463 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b3e2ea56-3525-4e78-a7c8-289a9a2406e6-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-8llfp\" (UID: \"b3e2ea56-3525-4e78-a7c8-289a9a2406e6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8llfp" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.411487 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/534e137b-1c8d-45fb-9c1b-6804913d309b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-r5ts6\" (UID: \"534e137b-1c8d-45fb-9c1b-6804913d309b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r5ts6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.411520 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/08b0652b-f30f-4c2c-bd1a-75b80a66b144-client-ca\") pod \"route-controller-manager-6576b87f9c-hzh5b\" (UID: \"08b0652b-f30f-4c2c-bd1a-75b80a66b144\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.411535 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.411838 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e814799e-e374-4ad8-96a2-46e4d5defc2f-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-ntvtt\" (UID: \"e814799e-e374-4ad8-96a2-46e4d5defc2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.411875 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41ae564b-da35-45eb-9a27-fbf23bacaf8f-serving-cert\") pod \"console-operator-58897d9998-rqftc\" (UID: \"41ae564b-da35-45eb-9a27-fbf23bacaf8f\") " pod="openshift-console-operator/console-operator-58897d9998-rqftc" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.411964 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08b0652b-f30f-4c2c-bd1a-75b80a66b144-config\") pod \"route-controller-manager-6576b87f9c-hzh5b\" (UID: \"08b0652b-f30f-4c2c-bd1a-75b80a66b144\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.412056 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f90213c-484b-4f9d-b064-363887589530-service-ca-bundle\") pod \"authentication-operator-69f744f599-q5d8b\" (UID: \"7f90213c-484b-4f9d-b064-363887589530\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q5d8b" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.412144 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/994432aa-8794-4818-94bd-ce28d97c7163-trusted-ca-bundle\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.412401 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f90213c-484b-4f9d-b064-363887589530-config\") pod \"authentication-operator-69f744f599-q5d8b\" (UID: \"7f90213c-484b-4f9d-b064-363887589530\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q5d8b" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.412671 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e7b5892a-8920-4cef-94a1-3736328e9a39-audit-policies\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.412698 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/534e137b-1c8d-45fb-9c1b-6804913d309b-serving-cert\") pod \"controller-manager-879f6c89f-r5ts6\" (UID: \"534e137b-1c8d-45fb-9c1b-6804913d309b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r5ts6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.412759 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e814799e-e374-4ad8-96a2-46e4d5defc2f-encryption-config\") pod \"apiserver-7bbb656c7d-ntvtt\" (UID: \"e814799e-e374-4ad8-96a2-46e4d5defc2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.413014 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6796a9bd-a428-4b0a-8723-d8a6a252de84-console-oauth-config\") pod \"console-f9d7485db-xvd4d\" (UID: \"6796a9bd-a428-4b0a-8723-d8a6a252de84\") " pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.413092 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/534e137b-1c8d-45fb-9c1b-6804913d309b-config\") pod \"controller-manager-879f6c89f-r5ts6\" (UID: \"534e137b-1c8d-45fb-9c1b-6804913d309b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r5ts6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.413417 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41ae564b-da35-45eb-9a27-fbf23bacaf8f-config\") pod \"console-operator-58897d9998-rqftc\" (UID: \"41ae564b-da35-45eb-9a27-fbf23bacaf8f\") " pod="openshift-console-operator/console-operator-58897d9998-rqftc" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.413720 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e814799e-e374-4ad8-96a2-46e4d5defc2f-serving-cert\") pod \"apiserver-7bbb656c7d-ntvtt\" (UID: \"e814799e-e374-4ad8-96a2-46e4d5defc2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.413950 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.414090 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.414397 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6796a9bd-a428-4b0a-8723-d8a6a252de84-console-serving-cert\") pod \"console-f9d7485db-xvd4d\" (UID: \"6796a9bd-a428-4b0a-8723-d8a6a252de84\") " pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.414569 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.414782 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/410014cb-6ba0-4610-acc5-38f3db8fcae2-machine-approver-tls\") pod \"machine-approver-56656f9798-l7z49\" (UID: \"410014cb-6ba0-4610-acc5-38f3db8fcae2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7z49" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.414937 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e814799e-e374-4ad8-96a2-46e4d5defc2f-etcd-client\") pod \"apiserver-7bbb656c7d-ntvtt\" (UID: \"e814799e-e374-4ad8-96a2-46e4d5defc2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.414992 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.415003 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.415161 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.416027 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/994432aa-8794-4818-94bd-ce28d97c7163-etcd-client\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.416017 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/18f99d77-1c2d-412b-93ae-1d1fc52f24ab-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-ztnsp\" (UID: \"18f99d77-1c2d-412b-93ae-1d1fc52f24ab\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ztnsp" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.416189 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e417e794-f6b2-4607-b2ce-3a737e8770b4-serving-cert\") pod \"openshift-config-operator-7777fb866f-fjv2r\" (UID: \"e417e794-f6b2-4607-b2ce-3a737e8770b4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fjv2r" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.416724 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08b0652b-f30f-4c2c-bd1a-75b80a66b144-serving-cert\") pod \"route-controller-manager-6576b87f9c-hzh5b\" (UID: \"08b0652b-f30f-4c2c-bd1a-75b80a66b144\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.417081 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f90213c-484b-4f9d-b064-363887589530-serving-cert\") pod \"authentication-operator-69f744f599-q5d8b\" (UID: \"7f90213c-484b-4f9d-b064-363887589530\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q5d8b" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.424616 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-b2l6v"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.425100 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-b2l6v" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.427042 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e9514f9-f14f-4409-8fef-b89a37de70a6-serving-cert\") pod \"etcd-operator-b45778765-9m2sk\" (UID: \"9e9514f9-f14f-4409-8fef-b89a37de70a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9m2sk" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.428534 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-b2l6v"] Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.433723 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.452938 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.461123 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/9e9514f9-f14f-4409-8fef-b89a37de70a6-etcd-service-ca\") pod \"etcd-operator-b45778765-9m2sk\" (UID: \"9e9514f9-f14f-4409-8fef-b89a37de70a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9m2sk" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.474268 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.481857 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e9514f9-f14f-4409-8fef-b89a37de70a6-config\") pod \"etcd-operator-b45778765-9m2sk\" (UID: \"9e9514f9-f14f-4409-8fef-b89a37de70a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9m2sk" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.507468 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2zdw\" (UniqueName: \"kubernetes.io/projected/a91fbd9b-8264-4ac4-b1bd-d62143b7a4d2-kube-api-access-f2zdw\") pod \"multus-admission-controller-857f4d67dd-8wt84\" (UID: \"a91fbd9b-8264-4ac4-b1bd-d62143b7a4d2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8wt84" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.507497 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3d1342fa-3bb4-4924-a36b-e370aec66cd3-certs\") pod \"machine-config-server-jgl4n\" (UID: \"3d1342fa-3bb4-4924-a36b-e370aec66cd3\") " pod="openshift-machine-config-operator/machine-config-server-jgl4n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.507517 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szjqc\" (UniqueName: \"kubernetes.io/projected/0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb-kube-api-access-szjqc\") pod \"catalog-operator-68c6474976-mkb95\" (UID: \"0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mkb95" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.507538 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d1d0d430-031f-448e-9763-69deb17abc4a-secret-volume\") pod \"collect-profiles-29398005-nnldq\" (UID: \"d1d0d430-031f-448e-9763-69deb17abc4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-nnldq" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.507559 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/528f6e87-f9e3-46f4-b40b-127b4092b8aa-signing-cabundle\") pod \"service-ca-9c57cc56f-ltkhr\" (UID: \"528f6e87-f9e3-46f4-b40b-127b4092b8aa\") " pod="openshift-service-ca/service-ca-9c57cc56f-ltkhr" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.507579 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/09d28a3a-5b25-4e31-90d6-17d9319cd11c-proxy-tls\") pod \"machine-config-controller-84d6567774-t2rfc\" (UID: \"09d28a3a-5b25-4e31-90d6-17d9319cd11c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t2rfc" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.507593 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9cjx\" (UniqueName: \"kubernetes.io/projected/35b44360-64aa-45b1-9141-15d5f69b384a-kube-api-access-z9cjx\") pod \"kube-storage-version-migrator-operator-b67b599dd-ggt5f\" (UID: \"35b44360-64aa-45b1-9141-15d5f69b384a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ggt5f" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.507611 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/337f295d-6c74-4850-bb14-125192de4385-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tgbrt\" (UID: \"337f295d-6c74-4850-bb14-125192de4385\") " pod="openshift-marketplace/marketplace-operator-79b997595-tgbrt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.507627 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba-service-ca-bundle\") pod \"router-default-5444994796-9gwbh\" (UID: \"a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba\") " pod="openshift-ingress/router-default-5444994796-9gwbh" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.507655 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a130a936-3db2-4455-bb63-fcea37eda82c-config\") pod \"kube-controller-manager-operator-78b949d7b-8xfrw\" (UID: \"a130a936-3db2-4455-bb63-fcea37eda82c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8xfrw" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.507673 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/ae746e9c-3187-4a3b-a439-b5ef25b37caf-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-2j2z7\" (UID: \"ae746e9c-3187-4a3b-a439-b5ef25b37caf\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2j2z7" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.507690 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/833dd52a-3337-431a-af83-d75240e6ab5c-apiservice-cert\") pod \"packageserver-d55dfcdfc-78zq2\" (UID: \"833dd52a-3337-431a-af83-d75240e6ab5c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-78zq2" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.507703 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba-metrics-certs\") pod \"router-default-5444994796-9gwbh\" (UID: \"a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba\") " pod="openshift-ingress/router-default-5444994796-9gwbh" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.507716 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3d1342fa-3bb4-4924-a36b-e370aec66cd3-node-bootstrap-token\") pod \"machine-config-server-jgl4n\" (UID: \"3d1342fa-3bb4-4924-a36b-e370aec66cd3\") " pod="openshift-machine-config-operator/machine-config-server-jgl4n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.507731 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba-stats-auth\") pod \"router-default-5444994796-9gwbh\" (UID: \"a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba\") " pod="openshift-ingress/router-default-5444994796-9gwbh" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.507747 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5hv8\" (UniqueName: \"kubernetes.io/projected/50e7a859-2c04-4eff-889d-633ced116f34-kube-api-access-l5hv8\") pod \"machine-config-operator-74547568cd-lktfh\" (UID: \"50e7a859-2c04-4eff-889d-633ced116f34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lktfh" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.507771 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzshk\" (UniqueName: \"kubernetes.io/projected/d1d0d430-031f-448e-9763-69deb17abc4a-kube-api-access-nzshk\") pod \"collect-profiles-29398005-nnldq\" (UID: \"d1d0d430-031f-448e-9763-69deb17abc4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-nnldq" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.507785 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a130a936-3db2-4455-bb63-fcea37eda82c-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-8xfrw\" (UID: \"a130a936-3db2-4455-bb63-fcea37eda82c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8xfrw" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.507802 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/833dd52a-3337-431a-af83-d75240e6ab5c-webhook-cert\") pod \"packageserver-d55dfcdfc-78zq2\" (UID: \"833dd52a-3337-431a-af83-d75240e6ab5c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-78zq2" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.507828 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a922bdd-f7be-41d2-8626-9352ec57d9ce-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zr8cm\" (UID: \"3a922bdd-f7be-41d2-8626-9352ec57d9ce\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zr8cm" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.507842 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a91fbd9b-8264-4ac4-b1bd-d62143b7a4d2-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-8wt84\" (UID: \"a91fbd9b-8264-4ac4-b1bd-d62143b7a4d2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8wt84" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.507857 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5658\" (UniqueName: \"kubernetes.io/projected/2879a16a-7ab6-413a-a8a3-ecfa7ffd4ac7-kube-api-access-w5658\") pod \"ingress-operator-5b745b69d9-dnz8k\" (UID: \"2879a16a-7ab6-413a-a8a3-ecfa7ffd4ac7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dnz8k" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.507884 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b7acf5d-cae8-429a-8661-ec3abcf056c0-config\") pod \"kube-apiserver-operator-766d6c64bb-xwb9l\" (UID: \"1b7acf5d-cae8-429a-8661-ec3abcf056c0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xwb9l" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.507902 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1b7acf5d-cae8-429a-8661-ec3abcf056c0-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-xwb9l\" (UID: \"1b7acf5d-cae8-429a-8661-ec3abcf056c0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xwb9l" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.507922 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/50e7a859-2c04-4eff-889d-633ced116f34-proxy-tls\") pod \"machine-config-operator-74547568cd-lktfh\" (UID: \"50e7a859-2c04-4eff-889d-633ced116f34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lktfh" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.507937 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/528f6e87-f9e3-46f4-b40b-127b4092b8aa-signing-key\") pod \"service-ca-9c57cc56f-ltkhr\" (UID: \"528f6e87-f9e3-46f4-b40b-127b4092b8aa\") " pod="openshift-service-ca/service-ca-9c57cc56f-ltkhr" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.507962 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znnk8\" (UniqueName: \"kubernetes.io/projected/266931fd-4205-4e31-9ded-9882a2a41921-kube-api-access-znnk8\") pod \"migrator-59844c95c7-9775t\" (UID: \"266931fd-4205-4e31-9ded-9882a2a41921\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9775t" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508003 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2879a16a-7ab6-413a-a8a3-ecfa7ffd4ac7-metrics-tls\") pod \"ingress-operator-5b745b69d9-dnz8k\" (UID: \"2879a16a-7ab6-413a-a8a3-ecfa7ffd4ac7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dnz8k" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508018 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb-srv-cert\") pod \"catalog-operator-68c6474976-mkb95\" (UID: \"0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mkb95" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508034 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9v4zr\" (UniqueName: \"kubernetes.io/projected/04abe9e5-6ffe-429d-9274-73d133fd0d52-kube-api-access-9v4zr\") pod \"package-server-manager-789f6589d5-r779q\" (UID: \"04abe9e5-6ffe-429d-9274-73d133fd0d52\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r779q" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508048 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbvd4\" (UniqueName: \"kubernetes.io/projected/a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba-kube-api-access-hbvd4\") pod \"router-default-5444994796-9gwbh\" (UID: \"a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba\") " pod="openshift-ingress/router-default-5444994796-9gwbh" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508066 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb-profile-collector-cert\") pod \"catalog-operator-68c6474976-mkb95\" (UID: \"0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mkb95" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508082 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n658h\" (UniqueName: \"kubernetes.io/projected/3d1342fa-3bb4-4924-a36b-e370aec66cd3-kube-api-access-n658h\") pod \"machine-config-server-jgl4n\" (UID: \"3d1342fa-3bb4-4924-a36b-e370aec66cd3\") " pod="openshift-machine-config-operator/machine-config-server-jgl4n" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508100 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k45sh\" (UniqueName: \"kubernetes.io/projected/09d28a3a-5b25-4e31-90d6-17d9319cd11c-kube-api-access-k45sh\") pod \"machine-config-controller-84d6567774-t2rfc\" (UID: \"09d28a3a-5b25-4e31-90d6-17d9319cd11c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t2rfc" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508115 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2879a16a-7ab6-413a-a8a3-ecfa7ffd4ac7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-dnz8k\" (UID: \"2879a16a-7ab6-413a-a8a3-ecfa7ffd4ac7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dnz8k" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508135 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b7acf5d-cae8-429a-8661-ec3abcf056c0-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-xwb9l\" (UID: \"1b7acf5d-cae8-429a-8661-ec3abcf056c0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xwb9l" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508149 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3a922bdd-f7be-41d2-8626-9352ec57d9ce-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zr8cm\" (UID: \"3a922bdd-f7be-41d2-8626-9352ec57d9ce\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zr8cm" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508163 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d1d0d430-031f-448e-9763-69deb17abc4a-config-volume\") pod \"collect-profiles-29398005-nnldq\" (UID: \"d1d0d430-031f-448e-9763-69deb17abc4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-nnldq" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508179 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4zsf\" (UniqueName: \"kubernetes.io/projected/ae746e9c-3187-4a3b-a439-b5ef25b37caf-kube-api-access-x4zsf\") pod \"control-plane-machine-set-operator-78cbb6b69f-2j2z7\" (UID: \"ae746e9c-3187-4a3b-a439-b5ef25b37caf\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2j2z7" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508201 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmdwc\" (UniqueName: \"kubernetes.io/projected/80d890b8-c797-4392-a48a-7894059147d4-kube-api-access-bmdwc\") pod \"openshift-controller-manager-operator-756b6f6bc6-qt4fl\" (UID: \"80d890b8-c797-4392-a48a-7894059147d4\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qt4fl" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508218 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/09d28a3a-5b25-4e31-90d6-17d9319cd11c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-t2rfc\" (UID: \"09d28a3a-5b25-4e31-90d6-17d9319cd11c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t2rfc" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508284 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/04abe9e5-6ffe-429d-9274-73d133fd0d52-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-r779q\" (UID: \"04abe9e5-6ffe-429d-9274-73d133fd0d52\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r779q" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508300 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a130a936-3db2-4455-bb63-fcea37eda82c-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-8xfrw\" (UID: \"a130a936-3db2-4455-bb63-fcea37eda82c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8xfrw" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508322 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a922bdd-f7be-41d2-8626-9352ec57d9ce-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zr8cm\" (UID: \"3a922bdd-f7be-41d2-8626-9352ec57d9ce\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zr8cm" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508362 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqq59\" (UniqueName: \"kubernetes.io/projected/337f295d-6c74-4850-bb14-125192de4385-kube-api-access-pqq59\") pod \"marketplace-operator-79b997595-tgbrt\" (UID: \"337f295d-6c74-4850-bb14-125192de4385\") " pod="openshift-marketplace/marketplace-operator-79b997595-tgbrt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508377 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80d890b8-c797-4392-a48a-7894059147d4-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-qt4fl\" (UID: \"80d890b8-c797-4392-a48a-7894059147d4\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qt4fl" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508391 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn86q\" (UniqueName: \"kubernetes.io/projected/833dd52a-3337-431a-af83-d75240e6ab5c-kube-api-access-fn86q\") pod \"packageserver-d55dfcdfc-78zq2\" (UID: \"833dd52a-3337-431a-af83-d75240e6ab5c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-78zq2" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508405 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/50e7a859-2c04-4eff-889d-633ced116f34-auth-proxy-config\") pod \"machine-config-operator-74547568cd-lktfh\" (UID: \"50e7a859-2c04-4eff-889d-633ced116f34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lktfh" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508446 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/833dd52a-3337-431a-af83-d75240e6ab5c-tmpfs\") pod \"packageserver-d55dfcdfc-78zq2\" (UID: \"833dd52a-3337-431a-af83-d75240e6ab5c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-78zq2" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508462 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35b44360-64aa-45b1-9141-15d5f69b384a-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ggt5f\" (UID: \"35b44360-64aa-45b1-9141-15d5f69b384a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ggt5f" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508475 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m24rv\" (UniqueName: \"kubernetes.io/projected/528f6e87-f9e3-46f4-b40b-127b4092b8aa-kube-api-access-m24rv\") pod \"service-ca-9c57cc56f-ltkhr\" (UID: \"528f6e87-f9e3-46f4-b40b-127b4092b8aa\") " pod="openshift-service-ca/service-ca-9c57cc56f-ltkhr" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508529 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/50e7a859-2c04-4eff-889d-633ced116f34-images\") pod \"machine-config-operator-74547568cd-lktfh\" (UID: \"50e7a859-2c04-4eff-889d-633ced116f34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lktfh" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508544 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/337f295d-6c74-4850-bb14-125192de4385-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tgbrt\" (UID: \"337f295d-6c74-4850-bb14-125192de4385\") " pod="openshift-marketplace/marketplace-operator-79b997595-tgbrt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508561 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2879a16a-7ab6-413a-a8a3-ecfa7ffd4ac7-trusted-ca\") pod \"ingress-operator-5b745b69d9-dnz8k\" (UID: \"2879a16a-7ab6-413a-a8a3-ecfa7ffd4ac7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dnz8k" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508615 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/35b44360-64aa-45b1-9141-15d5f69b384a-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ggt5f\" (UID: \"35b44360-64aa-45b1-9141-15d5f69b384a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ggt5f" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508653 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba-default-certificate\") pod \"router-default-5444994796-9gwbh\" (UID: \"a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba\") " pod="openshift-ingress/router-default-5444994796-9gwbh" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.508677 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80d890b8-c797-4392-a48a-7894059147d4-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-qt4fl\" (UID: \"80d890b8-c797-4392-a48a-7894059147d4\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qt4fl" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.509377 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/09d28a3a-5b25-4e31-90d6-17d9319cd11c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-t2rfc\" (UID: \"09d28a3a-5b25-4e31-90d6-17d9319cd11c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t2rfc" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.509596 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/833dd52a-3337-431a-af83-d75240e6ab5c-tmpfs\") pod \"packageserver-d55dfcdfc-78zq2\" (UID: \"833dd52a-3337-431a-af83-d75240e6ab5c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-78zq2" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.509637 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/50e7a859-2c04-4eff-889d-633ced116f34-auth-proxy-config\") pod \"machine-config-operator-74547568cd-lktfh\" (UID: \"50e7a859-2c04-4eff-889d-633ced116f34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lktfh" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.516593 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.533425 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.553544 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.561763 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/35b44360-64aa-45b1-9141-15d5f69b384a-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ggt5f\" (UID: \"35b44360-64aa-45b1-9141-15d5f69b384a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ggt5f" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.573877 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.580156 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35b44360-64aa-45b1-9141-15d5f69b384a-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ggt5f\" (UID: \"35b44360-64aa-45b1-9141-15d5f69b384a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ggt5f" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.593984 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.613728 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.633483 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.640860 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/09d28a3a-5b25-4e31-90d6-17d9319cd11c-proxy-tls\") pod \"machine-config-controller-84d6567774-t2rfc\" (UID: \"09d28a3a-5b25-4e31-90d6-17d9319cd11c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t2rfc" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.653130 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.673896 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.679342 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80d890b8-c797-4392-a48a-7894059147d4-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-qt4fl\" (UID: \"80d890b8-c797-4392-a48a-7894059147d4\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qt4fl" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.693921 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.701495 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80d890b8-c797-4392-a48a-7894059147d4-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-qt4fl\" (UID: \"80d890b8-c797-4392-a48a-7894059147d4\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qt4fl" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.713851 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.733673 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.753667 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.760660 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba-metrics-certs\") pod \"router-default-5444994796-9gwbh\" (UID: \"a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba\") " pod="openshift-ingress/router-default-5444994796-9gwbh" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.773849 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.779852 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba-stats-auth\") pod \"router-default-5444994796-9gwbh\" (UID: \"a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba\") " pod="openshift-ingress/router-default-5444994796-9gwbh" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.794125 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.798632 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba-service-ca-bundle\") pod \"router-default-5444994796-9gwbh\" (UID: \"a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba\") " pod="openshift-ingress/router-default-5444994796-9gwbh" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.813620 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.833636 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.841754 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba-default-certificate\") pod \"router-default-5444994796-9gwbh\" (UID: \"a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba\") " pod="openshift-ingress/router-default-5444994796-9gwbh" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.853772 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.873070 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.893353 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.914102 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.933349 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.941812 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/04abe9e5-6ffe-429d-9274-73d133fd0d52-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-r779q\" (UID: \"04abe9e5-6ffe-429d-9274-73d133fd0d52\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r779q" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.954220 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.973754 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 23 06:47:11 crc kubenswrapper[4559]: I1123 06:47:11.993697 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.013511 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.033500 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.053845 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.073479 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.081126 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/833dd52a-3337-431a-af83-d75240e6ab5c-webhook-cert\") pod \"packageserver-d55dfcdfc-78zq2\" (UID: \"833dd52a-3337-431a-af83-d75240e6ab5c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-78zq2" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.081184 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/833dd52a-3337-431a-af83-d75240e6ab5c-apiservice-cert\") pod \"packageserver-d55dfcdfc-78zq2\" (UID: \"833dd52a-3337-431a-af83-d75240e6ab5c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-78zq2" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.093487 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.113227 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.120295 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a91fbd9b-8264-4ac4-b1bd-d62143b7a4d2-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-8wt84\" (UID: \"a91fbd9b-8264-4ac4-b1bd-d62143b7a4d2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8wt84" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.133611 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.157655 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.161356 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2879a16a-7ab6-413a-a8a3-ecfa7ffd4ac7-trusted-ca\") pod \"ingress-operator-5b745b69d9-dnz8k\" (UID: \"2879a16a-7ab6-413a-a8a3-ecfa7ffd4ac7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dnz8k" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.174245 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.193178 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.213549 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.233729 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.241098 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2879a16a-7ab6-413a-a8a3-ecfa7ffd4ac7-metrics-tls\") pod \"ingress-operator-5b745b69d9-dnz8k\" (UID: \"2879a16a-7ab6-413a-a8a3-ecfa7ffd4ac7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dnz8k" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.253595 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.262143 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a922bdd-f7be-41d2-8626-9352ec57d9ce-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zr8cm\" (UID: \"3a922bdd-f7be-41d2-8626-9352ec57d9ce\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zr8cm" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.273139 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.279195 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a922bdd-f7be-41d2-8626-9352ec57d9ce-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zr8cm\" (UID: \"3a922bdd-f7be-41d2-8626-9352ec57d9ce\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zr8cm" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.292918 4559 request.go:700] Waited for 1.018533613s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager-operator/configmaps?fieldSelector=metadata.name%3Dkube-controller-manager-operator-config&limit=500&resourceVersion=0 Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.293636 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.299007 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a130a936-3db2-4455-bb63-fcea37eda82c-config\") pod \"kube-controller-manager-operator-78b949d7b-8xfrw\" (UID: \"a130a936-3db2-4455-bb63-fcea37eda82c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8xfrw" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.313510 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.333102 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.342045 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a130a936-3db2-4455-bb63-fcea37eda82c-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-8xfrw\" (UID: \"a130a936-3db2-4455-bb63-fcea37eda82c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8xfrw" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.353279 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.360012 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b7acf5d-cae8-429a-8661-ec3abcf056c0-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-xwb9l\" (UID: \"1b7acf5d-cae8-429a-8661-ec3abcf056c0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xwb9l" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.373294 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.394187 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.413262 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.418939 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b7acf5d-cae8-429a-8661-ec3abcf056c0-config\") pod \"kube-apiserver-operator-766d6c64bb-xwb9l\" (UID: \"1b7acf5d-cae8-429a-8661-ec3abcf056c0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xwb9l" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.433772 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.453830 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.474233 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.480079 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/50e7a859-2c04-4eff-889d-633ced116f34-proxy-tls\") pod \"machine-config-operator-74547568cd-lktfh\" (UID: \"50e7a859-2c04-4eff-889d-633ced116f34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lktfh" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.493717 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 23 06:47:12 crc kubenswrapper[4559]: E1123 06:47:12.507850 4559 secret.go:188] Couldn't get secret openshift-machine-api/control-plane-machine-set-operator-tls: failed to sync secret cache: timed out waiting for the condition Nov 23 06:47:12 crc kubenswrapper[4559]: E1123 06:47:12.507894 4559 configmap.go:193] Couldn't get configMap openshift-service-ca/signing-cabundle: failed to sync configmap cache: timed out waiting for the condition Nov 23 06:47:12 crc kubenswrapper[4559]: E1123 06:47:12.507907 4559 secret.go:188] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Nov 23 06:47:12 crc kubenswrapper[4559]: E1123 06:47:12.507924 4559 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Nov 23 06:47:12 crc kubenswrapper[4559]: E1123 06:47:12.507941 4559 secret.go:188] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Nov 23 06:47:12 crc kubenswrapper[4559]: E1123 06:47:12.507946 4559 secret.go:188] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: failed to sync secret cache: timed out waiting for the condition Nov 23 06:47:12 crc kubenswrapper[4559]: E1123 06:47:12.507911 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ae746e9c-3187-4a3b-a439-b5ef25b37caf-control-plane-machine-set-operator-tls podName:ae746e9c-3187-4a3b-a439-b5ef25b37caf nodeName:}" failed. No retries permitted until 2025-11-23 06:47:13.007895319 +0000 UTC m=+135.029880933 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "control-plane-machine-set-operator-tls" (UniqueName: "kubernetes.io/secret/ae746e9c-3187-4a3b-a439-b5ef25b37caf-control-plane-machine-set-operator-tls") pod "control-plane-machine-set-operator-78cbb6b69f-2j2z7" (UID: "ae746e9c-3187-4a3b-a439-b5ef25b37caf") : failed to sync secret cache: timed out waiting for the condition Nov 23 06:47:12 crc kubenswrapper[4559]: E1123 06:47:12.507983 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/528f6e87-f9e3-46f4-b40b-127b4092b8aa-signing-cabundle podName:528f6e87-f9e3-46f4-b40b-127b4092b8aa nodeName:}" failed. No retries permitted until 2025-11-23 06:47:13.007972095 +0000 UTC m=+135.029957710 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-cabundle" (UniqueName: "kubernetes.io/configmap/528f6e87-f9e3-46f4-b40b-127b4092b8aa-signing-cabundle") pod "service-ca-9c57cc56f-ltkhr" (UID: "528f6e87-f9e3-46f4-b40b-127b4092b8aa") : failed to sync configmap cache: timed out waiting for the condition Nov 23 06:47:12 crc kubenswrapper[4559]: E1123 06:47:12.507995 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3d1342fa-3bb4-4924-a36b-e370aec66cd3-node-bootstrap-token podName:3d1342fa-3bb4-4924-a36b-e370aec66cd3 nodeName:}" failed. No retries permitted until 2025-11-23 06:47:13.007989788 +0000 UTC m=+135.029975402 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/3d1342fa-3bb4-4924-a36b-e370aec66cd3-node-bootstrap-token") pod "machine-config-server-jgl4n" (UID: "3d1342fa-3bb4-4924-a36b-e370aec66cd3") : failed to sync secret cache: timed out waiting for the condition Nov 23 06:47:12 crc kubenswrapper[4559]: E1123 06:47:12.508006 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3d1342fa-3bb4-4924-a36b-e370aec66cd3-certs podName:3d1342fa-3bb4-4924-a36b-e370aec66cd3 nodeName:}" failed. No retries permitted until 2025-11-23 06:47:13.00800081 +0000 UTC m=+135.029986424 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/3d1342fa-3bb4-4924-a36b-e370aec66cd3-certs") pod "machine-config-server-jgl4n" (UID: "3d1342fa-3bb4-4924-a36b-e370aec66cd3") : failed to sync secret cache: timed out waiting for the condition Nov 23 06:47:12 crc kubenswrapper[4559]: E1123 06:47:12.508015 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/337f295d-6c74-4850-bb14-125192de4385-marketplace-operator-metrics podName:337f295d-6c74-4850-bb14-125192de4385 nodeName:}" failed. No retries permitted until 2025-11-23 06:47:13.008010528 +0000 UTC m=+135.029996143 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/337f295d-6c74-4850-bb14-125192de4385-marketplace-operator-metrics") pod "marketplace-operator-79b997595-tgbrt" (UID: "337f295d-6c74-4850-bb14-125192de4385") : failed to sync secret cache: timed out waiting for the condition Nov 23 06:47:12 crc kubenswrapper[4559]: E1123 06:47:12.508026 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d1d0d430-031f-448e-9763-69deb17abc4a-secret-volume podName:d1d0d430-031f-448e-9763-69deb17abc4a nodeName:}" failed. No retries permitted until 2025-11-23 06:47:13.008021078 +0000 UTC m=+135.030006692 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-volume" (UniqueName: "kubernetes.io/secret/d1d0d430-031f-448e-9763-69deb17abc4a-secret-volume") pod "collect-profiles-29398005-nnldq" (UID: "d1d0d430-031f-448e-9763-69deb17abc4a") : failed to sync secret cache: timed out waiting for the condition Nov 23 06:47:12 crc kubenswrapper[4559]: E1123 06:47:12.508031 4559 secret.go:188] Couldn't get secret openshift-service-ca/signing-key: failed to sync secret cache: timed out waiting for the condition Nov 23 06:47:12 crc kubenswrapper[4559]: E1123 06:47:12.508079 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/528f6e87-f9e3-46f4-b40b-127b4092b8aa-signing-key podName:528f6e87-f9e3-46f4-b40b-127b4092b8aa nodeName:}" failed. No retries permitted until 2025-11-23 06:47:13.008063969 +0000 UTC m=+135.030049583 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-key" (UniqueName: "kubernetes.io/secret/528f6e87-f9e3-46f4-b40b-127b4092b8aa-signing-key") pod "service-ca-9c57cc56f-ltkhr" (UID: "528f6e87-f9e3-46f4-b40b-127b4092b8aa") : failed to sync secret cache: timed out waiting for the condition Nov 23 06:47:12 crc kubenswrapper[4559]: E1123 06:47:12.509098 4559 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Nov 23 06:47:12 crc kubenswrapper[4559]: E1123 06:47:12.509113 4559 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/catalog-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 23 06:47:12 crc kubenswrapper[4559]: E1123 06:47:12.509098 4559 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Nov 23 06:47:12 crc kubenswrapper[4559]: E1123 06:47:12.509140 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d1d0d430-031f-448e-9763-69deb17abc4a-config-volume podName:d1d0d430-031f-448e-9763-69deb17abc4a nodeName:}" failed. No retries permitted until 2025-11-23 06:47:13.009127388 +0000 UTC m=+135.031113002 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/d1d0d430-031f-448e-9763-69deb17abc4a-config-volume") pod "collect-profiles-29398005-nnldq" (UID: "d1d0d430-031f-448e-9763-69deb17abc4a") : failed to sync configmap cache: timed out waiting for the condition Nov 23 06:47:12 crc kubenswrapper[4559]: E1123 06:47:12.509227 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb-srv-cert podName:0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb nodeName:}" failed. No retries permitted until 2025-11-23 06:47:13.009207631 +0000 UTC m=+135.031193245 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb-srv-cert") pod "catalog-operator-68c6474976-mkb95" (UID: "0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb") : failed to sync secret cache: timed out waiting for the condition Nov 23 06:47:12 crc kubenswrapper[4559]: E1123 06:47:12.509242 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb-profile-collector-cert podName:0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb nodeName:}" failed. No retries permitted until 2025-11-23 06:47:13.009234632 +0000 UTC m=+135.031220246 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "profile-collector-cert" (UniqueName: "kubernetes.io/secret/0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb-profile-collector-cert") pod "catalog-operator-68c6474976-mkb95" (UID: "0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb") : failed to sync secret cache: timed out waiting for the condition Nov 23 06:47:12 crc kubenswrapper[4559]: E1123 06:47:12.510222 4559 configmap.go:193] Couldn't get configMap openshift-marketplace/marketplace-trusted-ca: failed to sync configmap cache: timed out waiting for the condition Nov 23 06:47:12 crc kubenswrapper[4559]: E1123 06:47:12.510254 4559 configmap.go:193] Couldn't get configMap openshift-machine-config-operator/machine-config-operator-images: failed to sync configmap cache: timed out waiting for the condition Nov 23 06:47:12 crc kubenswrapper[4559]: E1123 06:47:12.510265 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/337f295d-6c74-4850-bb14-125192de4385-marketplace-trusted-ca podName:337f295d-6c74-4850-bb14-125192de4385 nodeName:}" failed. No retries permitted until 2025-11-23 06:47:13.010255791 +0000 UTC m=+135.032241406 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-trusted-ca" (UniqueName: "kubernetes.io/configmap/337f295d-6c74-4850-bb14-125192de4385-marketplace-trusted-ca") pod "marketplace-operator-79b997595-tgbrt" (UID: "337f295d-6c74-4850-bb14-125192de4385") : failed to sync configmap cache: timed out waiting for the condition Nov 23 06:47:12 crc kubenswrapper[4559]: E1123 06:47:12.510291 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/50e7a859-2c04-4eff-889d-633ced116f34-images podName:50e7a859-2c04-4eff-889d-633ced116f34 nodeName:}" failed. No retries permitted until 2025-11-23 06:47:13.010280218 +0000 UTC m=+135.032265832 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/50e7a859-2c04-4eff-889d-633ced116f34-images") pod "machine-config-operator-74547568cd-lktfh" (UID: "50e7a859-2c04-4eff-889d-633ced116f34") : failed to sync configmap cache: timed out waiting for the condition Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.513698 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.533848 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.553899 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.573315 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.593772 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.613740 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.633531 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.653721 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.673627 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.694132 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.713408 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.738442 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.753780 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.773825 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.794123 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.818173 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.833363 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.854034 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.873310 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.893883 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.913968 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.933821 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.953712 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.973078 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 23 06:47:12 crc kubenswrapper[4559]: I1123 06:47:12.993979 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.013363 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.026941 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d1d0d430-031f-448e-9763-69deb17abc4a-secret-volume\") pod \"collect-profiles-29398005-nnldq\" (UID: \"d1d0d430-031f-448e-9763-69deb17abc4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-nnldq" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.026994 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/528f6e87-f9e3-46f4-b40b-127b4092b8aa-signing-cabundle\") pod \"service-ca-9c57cc56f-ltkhr\" (UID: \"528f6e87-f9e3-46f4-b40b-127b4092b8aa\") " pod="openshift-service-ca/service-ca-9c57cc56f-ltkhr" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.027027 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/337f295d-6c74-4850-bb14-125192de4385-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tgbrt\" (UID: \"337f295d-6c74-4850-bb14-125192de4385\") " pod="openshift-marketplace/marketplace-operator-79b997595-tgbrt" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.027069 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/ae746e9c-3187-4a3b-a439-b5ef25b37caf-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-2j2z7\" (UID: \"ae746e9c-3187-4a3b-a439-b5ef25b37caf\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2j2z7" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.027089 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3d1342fa-3bb4-4924-a36b-e370aec66cd3-node-bootstrap-token\") pod \"machine-config-server-jgl4n\" (UID: \"3d1342fa-3bb4-4924-a36b-e370aec66cd3\") " pod="openshift-machine-config-operator/machine-config-server-jgl4n" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.027163 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/528f6e87-f9e3-46f4-b40b-127b4092b8aa-signing-key\") pod \"service-ca-9c57cc56f-ltkhr\" (UID: \"528f6e87-f9e3-46f4-b40b-127b4092b8aa\") " pod="openshift-service-ca/service-ca-9c57cc56f-ltkhr" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.027236 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb-srv-cert\") pod \"catalog-operator-68c6474976-mkb95\" (UID: \"0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mkb95" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.027279 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb-profile-collector-cert\") pod \"catalog-operator-68c6474976-mkb95\" (UID: \"0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mkb95" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.027743 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d1d0d430-031f-448e-9763-69deb17abc4a-config-volume\") pod \"collect-profiles-29398005-nnldq\" (UID: \"d1d0d430-031f-448e-9763-69deb17abc4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-nnldq" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.027854 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/50e7a859-2c04-4eff-889d-633ced116f34-images\") pod \"machine-config-operator-74547568cd-lktfh\" (UID: \"50e7a859-2c04-4eff-889d-633ced116f34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lktfh" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.027875 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/337f295d-6c74-4850-bb14-125192de4385-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tgbrt\" (UID: \"337f295d-6c74-4850-bb14-125192de4385\") " pod="openshift-marketplace/marketplace-operator-79b997595-tgbrt" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.027964 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3d1342fa-3bb4-4924-a36b-e370aec66cd3-certs\") pod \"machine-config-server-jgl4n\" (UID: \"3d1342fa-3bb4-4924-a36b-e370aec66cd3\") " pod="openshift-machine-config-operator/machine-config-server-jgl4n" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.028277 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/528f6e87-f9e3-46f4-b40b-127b4092b8aa-signing-cabundle\") pod \"service-ca-9c57cc56f-ltkhr\" (UID: \"528f6e87-f9e3-46f4-b40b-127b4092b8aa\") " pod="openshift-service-ca/service-ca-9c57cc56f-ltkhr" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.028921 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/337f295d-6c74-4850-bb14-125192de4385-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tgbrt\" (UID: \"337f295d-6c74-4850-bb14-125192de4385\") " pod="openshift-marketplace/marketplace-operator-79b997595-tgbrt" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.029210 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d1d0d430-031f-448e-9763-69deb17abc4a-config-volume\") pod \"collect-profiles-29398005-nnldq\" (UID: \"d1d0d430-031f-448e-9763-69deb17abc4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-nnldq" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.029367 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/50e7a859-2c04-4eff-889d-633ced116f34-images\") pod \"machine-config-operator-74547568cd-lktfh\" (UID: \"50e7a859-2c04-4eff-889d-633ced116f34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lktfh" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.029725 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb-srv-cert\") pod \"catalog-operator-68c6474976-mkb95\" (UID: \"0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mkb95" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.029970 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/ae746e9c-3187-4a3b-a439-b5ef25b37caf-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-2j2z7\" (UID: \"ae746e9c-3187-4a3b-a439-b5ef25b37caf\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2j2z7" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.030144 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d1d0d430-031f-448e-9763-69deb17abc4a-secret-volume\") pod \"collect-profiles-29398005-nnldq\" (UID: \"d1d0d430-031f-448e-9763-69deb17abc4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-nnldq" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.030416 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/337f295d-6c74-4850-bb14-125192de4385-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tgbrt\" (UID: \"337f295d-6c74-4850-bb14-125192de4385\") " pod="openshift-marketplace/marketplace-operator-79b997595-tgbrt" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.030571 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/528f6e87-f9e3-46f4-b40b-127b4092b8aa-signing-key\") pod \"service-ca-9c57cc56f-ltkhr\" (UID: \"528f6e87-f9e3-46f4-b40b-127b4092b8aa\") " pod="openshift-service-ca/service-ca-9c57cc56f-ltkhr" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.030677 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb-profile-collector-cert\") pod \"catalog-operator-68c6474976-mkb95\" (UID: \"0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mkb95" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.031225 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3d1342fa-3bb4-4924-a36b-e370aec66cd3-certs\") pod \"machine-config-server-jgl4n\" (UID: \"3d1342fa-3bb4-4924-a36b-e370aec66cd3\") " pod="openshift-machine-config-operator/machine-config-server-jgl4n" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.034028 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.039755 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3d1342fa-3bb4-4924-a36b-e370aec66cd3-node-bootstrap-token\") pod \"machine-config-server-jgl4n\" (UID: \"3d1342fa-3bb4-4924-a36b-e370aec66cd3\") " pod="openshift-machine-config-operator/machine-config-server-jgl4n" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.073205 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.093573 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.113330 4559 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.133430 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.153474 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.173635 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.204438 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8xcm\" (UniqueName: \"kubernetes.io/projected/9e9514f9-f14f-4409-8fef-b89a37de70a6-kube-api-access-f8xcm\") pod \"etcd-operator-b45778765-9m2sk\" (UID: \"9e9514f9-f14f-4409-8fef-b89a37de70a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9m2sk" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.223726 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z57mp\" (UniqueName: \"kubernetes.io/projected/e7b5892a-8920-4cef-94a1-3736328e9a39-kube-api-access-z57mp\") pod \"oauth-openshift-558db77b4-kg77n\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.244591 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l6xn\" (UniqueName: \"kubernetes.io/projected/e814799e-e374-4ad8-96a2-46e4d5defc2f-kube-api-access-6l6xn\") pod \"apiserver-7bbb656c7d-ntvtt\" (UID: \"e814799e-e374-4ad8-96a2-46e4d5defc2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.263638 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk4b8\" (UniqueName: \"kubernetes.io/projected/410014cb-6ba0-4610-acc5-38f3db8fcae2-kube-api-access-kk4b8\") pod \"machine-approver-56656f9798-l7z49\" (UID: \"410014cb-6ba0-4610-acc5-38f3db8fcae2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7z49" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.271679 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7z49" Nov 23 06:47:13 crc kubenswrapper[4559]: W1123 06:47:13.281445 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod410014cb_6ba0_4610_acc5_38f3db8fcae2.slice/crio-684d829615ee64b9d61b9d416fbae1d578a50636d3e23524652699cf5d2fe7d0 WatchSource:0}: Error finding container 684d829615ee64b9d61b9d416fbae1d578a50636d3e23524652699cf5d2fe7d0: Status 404 returned error can't find the container with id 684d829615ee64b9d61b9d416fbae1d578a50636d3e23524652699cf5d2fe7d0 Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.284474 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h642b\" (UniqueName: \"kubernetes.io/projected/08b0652b-f30f-4c2c-bd1a-75b80a66b144-kube-api-access-h642b\") pod \"route-controller-manager-6576b87f9c-hzh5b\" (UID: \"08b0652b-f30f-4c2c-bd1a-75b80a66b144\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.303983 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwb8n\" (UniqueName: \"kubernetes.io/projected/b3e2ea56-3525-4e78-a7c8-289a9a2406e6-kube-api-access-bwb8n\") pod \"cluster-image-registry-operator-dc59b4c8b-8llfp\" (UID: \"b3e2ea56-3525-4e78-a7c8-289a9a2406e6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8llfp" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.312227 4559 request.go:700] Waited for 1.903347984s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-cluster-samples-operator/serviceaccounts/cluster-samples-operator/token Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.324839 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxm7f\" (UniqueName: \"kubernetes.io/projected/58f094f8-8c16-4929-ae94-e52241c5b7d1-kube-api-access-qxm7f\") pod \"cluster-samples-operator-665b6dd947-j665r\" (UID: \"58f094f8-8c16-4929-ae94-e52241c5b7d1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j665r" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.344212 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6mxz\" (UniqueName: \"kubernetes.io/projected/41ae564b-da35-45eb-9a27-fbf23bacaf8f-kube-api-access-b6mxz\") pod \"console-operator-58897d9998-rqftc\" (UID: \"41ae564b-da35-45eb-9a27-fbf23bacaf8f\") " pod="openshift-console-operator/console-operator-58897d9998-rqftc" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.360429 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.367032 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zph27\" (UniqueName: \"kubernetes.io/projected/4316762b-b889-41c0-aa31-f874bc0fa3fe-kube-api-access-zph27\") pod \"downloads-7954f5f757-p9rps\" (UID: \"4316762b-b889-41c0-aa31-f874bc0fa3fe\") " pod="openshift-console/downloads-7954f5f757-p9rps" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.385685 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfjqs\" (UniqueName: \"kubernetes.io/projected/3a883895-9775-4c18-962e-be4df71043c8-kube-api-access-sfjqs\") pod \"openshift-apiserver-operator-796bbdcf4f-q79mk\" (UID: \"3a883895-9775-4c18-962e-be4df71043c8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q79mk" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.405115 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8j9h\" (UniqueName: \"kubernetes.io/projected/16e9fcf3-adc7-4af9-a625-125ff496cfcf-kube-api-access-t8j9h\") pod \"dns-operator-744455d44c-dlj2h\" (UID: \"16e9fcf3-adc7-4af9-a625-125ff496cfcf\") " pod="openshift-dns-operator/dns-operator-744455d44c-dlj2h" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.428944 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.431054 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnflg\" (UniqueName: \"kubernetes.io/projected/994432aa-8794-4818-94bd-ce28d97c7163-kube-api-access-qnflg\") pod \"apiserver-76f77b778f-5kfl6\" (UID: \"994432aa-8794-4818-94bd-ce28d97c7163\") " pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.439710 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.452092 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndlbq\" (UniqueName: \"kubernetes.io/projected/e417e794-f6b2-4607-b2ce-3a737e8770b4-kube-api-access-ndlbq\") pod \"openshift-config-operator-7777fb866f-fjv2r\" (UID: \"e417e794-f6b2-4607-b2ce-3a737e8770b4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fjv2r" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.452301 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j665r" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.457894 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-p9rps" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.461438 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-rqftc" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.464825 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b3e2ea56-3525-4e78-a7c8-289a9a2406e6-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-8llfp\" (UID: \"b3e2ea56-3525-4e78-a7c8-289a9a2406e6\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8llfp" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.466606 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8llfp" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.472459 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-dlj2h" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.477748 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-9m2sk" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.484735 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdt5m\" (UniqueName: \"kubernetes.io/projected/18f99d77-1c2d-412b-93ae-1d1fc52f24ab-kube-api-access-vdt5m\") pod \"machine-api-operator-5694c8668f-ztnsp\" (UID: \"18f99d77-1c2d-412b-93ae-1d1fc52f24ab\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ztnsp" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.512465 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fjff\" (UniqueName: \"kubernetes.io/projected/6796a9bd-a428-4b0a-8723-d8a6a252de84-kube-api-access-8fjff\") pod \"console-f9d7485db-xvd4d\" (UID: \"6796a9bd-a428-4b0a-8723-d8a6a252de84\") " pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.527045 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8hn2\" (UniqueName: \"kubernetes.io/projected/7f90213c-484b-4f9d-b064-363887589530-kube-api-access-g8hn2\") pod \"authentication-operator-69f744f599-q5d8b\" (UID: \"7f90213c-484b-4f9d-b064-363887589530\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q5d8b" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.550876 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hcm8\" (UniqueName: \"kubernetes.io/projected/534e137b-1c8d-45fb-9c1b-6804913d309b-kube-api-access-8hcm8\") pod \"controller-manager-879f6c89f-r5ts6\" (UID: \"534e137b-1c8d-45fb-9c1b-6804913d309b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r5ts6" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.553897 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.573965 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.578390 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-kg77n"] Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.596362 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.613856 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.624887 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-ztnsp" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.633716 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.649275 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-r5ts6" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.665978 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2zdw\" (UniqueName: \"kubernetes.io/projected/a91fbd9b-8264-4ac4-b1bd-d62143b7a4d2-kube-api-access-f2zdw\") pod \"multus-admission-controller-857f4d67dd-8wt84\" (UID: \"a91fbd9b-8264-4ac4-b1bd-d62143b7a4d2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8wt84" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.670089 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" event={"ID":"e7b5892a-8920-4cef-94a1-3736328e9a39","Type":"ContainerStarted","Data":"dae9c05d7f770dec3301fbffab75de3359ae8a0a030639f242ee6c49d68826b6"} Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.672058 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7z49" event={"ID":"410014cb-6ba0-4610-acc5-38f3db8fcae2","Type":"ContainerStarted","Data":"60b4167ff048f83c3fcb7cdf49dfaa7f696f8a7d89f708062fa37d8b57364783"} Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.672091 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7z49" event={"ID":"410014cb-6ba0-4610-acc5-38f3db8fcae2","Type":"ContainerStarted","Data":"0afbf73a3bbd3783b4f133c2ab77d38b2ce49467a17b5ffac504dd8218eec169"} Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.672103 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7z49" event={"ID":"410014cb-6ba0-4610-acc5-38f3db8fcae2","Type":"ContainerStarted","Data":"684d829615ee64b9d61b9d416fbae1d578a50636d3e23524652699cf5d2fe7d0"} Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.674620 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q79mk" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.685527 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9cjx\" (UniqueName: \"kubernetes.io/projected/35b44360-64aa-45b1-9141-15d5f69b384a-kube-api-access-z9cjx\") pod \"kube-storage-version-migrator-operator-b67b599dd-ggt5f\" (UID: \"35b44360-64aa-45b1-9141-15d5f69b384a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ggt5f" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.695114 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-q5d8b" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.700417 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt"] Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.707768 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szjqc\" (UniqueName: \"kubernetes.io/projected/0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb-kube-api-access-szjqc\") pod \"catalog-operator-68c6474976-mkb95\" (UID: \"0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mkb95" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.709808 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-8wt84" Nov 23 06:47:13 crc kubenswrapper[4559]: W1123 06:47:13.709851 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode814799e_e374_4ad8_96a2_46e4d5defc2f.slice/crio-8dddf16b1e7b7e259df778cf1b48c434c43cb8672dcad0e7dcf6e855c0c1214a WatchSource:0}: Error finding container 8dddf16b1e7b7e259df778cf1b48c434c43cb8672dcad0e7dcf6e855c0c1214a: Status 404 returned error can't find the container with id 8dddf16b1e7b7e259df778cf1b48c434c43cb8672dcad0e7dcf6e855c0c1214a Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.728026 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5hv8\" (UniqueName: \"kubernetes.io/projected/50e7a859-2c04-4eff-889d-633ced116f34-kube-api-access-l5hv8\") pod \"machine-config-operator-74547568cd-lktfh\" (UID: \"50e7a859-2c04-4eff-889d-633ced116f34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lktfh" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.735269 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.748690 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fjv2r" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.751864 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzshk\" (UniqueName: \"kubernetes.io/projected/d1d0d430-031f-448e-9763-69deb17abc4a-kube-api-access-nzshk\") pod \"collect-profiles-29398005-nnldq\" (UID: \"d1d0d430-031f-448e-9763-69deb17abc4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-nnldq" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.769076 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a130a936-3db2-4455-bb63-fcea37eda82c-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-8xfrw\" (UID: \"a130a936-3db2-4455-bb63-fcea37eda82c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8xfrw" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.784599 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ggt5f" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.786658 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5658\" (UniqueName: \"kubernetes.io/projected/2879a16a-7ab6-413a-a8a3-ecfa7ffd4ac7-kube-api-access-w5658\") pod \"ingress-operator-5b745b69d9-dnz8k\" (UID: \"2879a16a-7ab6-413a-a8a3-ecfa7ffd4ac7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dnz8k" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.805375 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1b7acf5d-cae8-429a-8661-ec3abcf056c0-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-xwb9l\" (UID: \"1b7acf5d-cae8-429a-8661-ec3abcf056c0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xwb9l" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.828180 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znnk8\" (UniqueName: \"kubernetes.io/projected/266931fd-4205-4e31-9ded-9882a2a41921-kube-api-access-znnk8\") pod \"migrator-59844c95c7-9775t\" (UID: \"266931fd-4205-4e31-9ded-9882a2a41921\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9775t" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.848273 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9v4zr\" (UniqueName: \"kubernetes.io/projected/04abe9e5-6ffe-429d-9274-73d133fd0d52-kube-api-access-9v4zr\") pod \"package-server-manager-789f6589d5-r779q\" (UID: \"04abe9e5-6ffe-429d-9274-73d133fd0d52\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r779q" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.848393 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j665r"] Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.849820 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b"] Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.855268 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8xfrw" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.869822 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbvd4\" (UniqueName: \"kubernetes.io/projected/a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba-kube-api-access-hbvd4\") pod \"router-default-5444994796-9gwbh\" (UID: \"a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba\") " pod="openshift-ingress/router-default-5444994796-9gwbh" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.876936 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xwb9l" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.882063 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r5ts6"] Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.882101 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q79mk"] Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.884415 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9775t" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.894953 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lktfh" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.899488 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n658h\" (UniqueName: \"kubernetes.io/projected/3d1342fa-3bb4-4924-a36b-e370aec66cd3-kube-api-access-n658h\") pod \"machine-config-server-jgl4n\" (UID: \"3d1342fa-3bb4-4924-a36b-e370aec66cd3\") " pod="openshift-machine-config-operator/machine-config-server-jgl4n" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.902202 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-p9rps"] Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.903259 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-rqftc"] Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.914346 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-nnldq" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.915202 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-9m2sk"] Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.919929 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8llfp"] Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.923513 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-dlj2h"] Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.924707 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k45sh\" (UniqueName: \"kubernetes.io/projected/09d28a3a-5b25-4e31-90d6-17d9319cd11c-kube-api-access-k45sh\") pod \"machine-config-controller-84d6567774-t2rfc\" (UID: \"09d28a3a-5b25-4e31-90d6-17d9319cd11c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t2rfc" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.927367 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mkb95" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.929145 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2879a16a-7ab6-413a-a8a3-ecfa7ffd4ac7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-dnz8k\" (UID: \"2879a16a-7ab6-413a-a8a3-ecfa7ffd4ac7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dnz8k" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.939099 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-jgl4n" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.952463 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3a922bdd-f7be-41d2-8626-9352ec57d9ce-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zr8cm\" (UID: \"3a922bdd-f7be-41d2-8626-9352ec57d9ce\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zr8cm" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.983428 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4zsf\" (UniqueName: \"kubernetes.io/projected/ae746e9c-3187-4a3b-a439-b5ef25b37caf-kube-api-access-x4zsf\") pod \"control-plane-machine-set-operator-78cbb6b69f-2j2z7\" (UID: \"ae746e9c-3187-4a3b-a439-b5ef25b37caf\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2j2z7" Nov 23 06:47:13 crc kubenswrapper[4559]: I1123 06:47:13.984578 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-xvd4d"] Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.018887 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dnz8k" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.028214 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn86q\" (UniqueName: \"kubernetes.io/projected/833dd52a-3337-431a-af83-d75240e6ab5c-kube-api-access-fn86q\") pod \"packageserver-d55dfcdfc-78zq2\" (UID: \"833dd52a-3337-431a-af83-d75240e6ab5c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-78zq2" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.030810 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmdwc\" (UniqueName: \"kubernetes.io/projected/80d890b8-c797-4392-a48a-7894059147d4-kube-api-access-bmdwc\") pod \"openshift-controller-manager-operator-756b6f6bc6-qt4fl\" (UID: \"80d890b8-c797-4392-a48a-7894059147d4\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qt4fl" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.030953 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqq59\" (UniqueName: \"kubernetes.io/projected/337f295d-6c74-4850-bb14-125192de4385-kube-api-access-pqq59\") pod \"marketplace-operator-79b997595-tgbrt\" (UID: \"337f295d-6c74-4850-bb14-125192de4385\") " pod="openshift-marketplace/marketplace-operator-79b997595-tgbrt" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.045113 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-5kfl6"] Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.050615 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-ztnsp"] Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.055579 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m24rv\" (UniqueName: \"kubernetes.io/projected/528f6e87-f9e3-46f4-b40b-127b4092b8aa-kube-api-access-m24rv\") pod \"service-ca-9c57cc56f-ltkhr\" (UID: \"528f6e87-f9e3-46f4-b40b-127b4092b8aa\") " pod="openshift-service-ca/service-ca-9c57cc56f-ltkhr" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.065144 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-fjv2r"] Nov 23 06:47:14 crc kubenswrapper[4559]: W1123 06:47:14.083936 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6796a9bd_a428_4b0a_8723_d8a6a252de84.slice/crio-e813441ff551059a86acca42beb61e7154ff1ea903a964f49529e56b8afc01fc WatchSource:0}: Error finding container e813441ff551059a86acca42beb61e7154ff1ea903a964f49529e56b8afc01fc: Status 404 returned error can't find the container with id e813441ff551059a86acca42beb61e7154ff1ea903a964f49529e56b8afc01fc Nov 23 06:47:14 crc kubenswrapper[4559]: W1123 06:47:14.086551 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode417e794_f6b2_4607_b2ce_3a737e8770b4.slice/crio-0e0f701368bcb381d71ad9a56ff021e108914328946fbc315a9e12fcb4010f1b WatchSource:0}: Error finding container 0e0f701368bcb381d71ad9a56ff021e108914328946fbc315a9e12fcb4010f1b: Status 404 returned error can't find the container with id 0e0f701368bcb381d71ad9a56ff021e108914328946fbc315a9e12fcb4010f1b Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.090571 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t2rfc" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.096840 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qt4fl" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.103710 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ggt5f"] Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.104138 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-9gwbh" Nov 23 06:47:14 crc kubenswrapper[4559]: W1123 06:47:14.104755 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod994432aa_8794_4818_94bd_ce28d97c7163.slice/crio-8962345c29b078925ab6d68c9a56651f2587982e22f1156c34f40a8ad0aabf5a WatchSource:0}: Error finding container 8962345c29b078925ab6d68c9a56651f2587982e22f1156c34f40a8ad0aabf5a: Status 404 returned error can't find the container with id 8962345c29b078925ab6d68c9a56651f2587982e22f1156c34f40a8ad0aabf5a Nov 23 06:47:14 crc kubenswrapper[4559]: W1123 06:47:14.116729 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18f99d77_1c2d_412b_93ae_1d1fc52f24ab.slice/crio-e6006e8a4df21bafcfc14739937a750bddb5536b03f6bcb88bc8d80ee7ee2b2c WatchSource:0}: Error finding container e6006e8a4df21bafcfc14739937a750bddb5536b03f6bcb88bc8d80ee7ee2b2c: Status 404 returned error can't find the container with id e6006e8a4df21bafcfc14739937a750bddb5536b03f6bcb88bc8d80ee7ee2b2c Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.139561 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8xfrw"] Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.144965 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rg7c6\" (UniqueName: \"kubernetes.io/projected/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-kube-api-access-rg7c6\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.145007 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-registry-tls\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.145024 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4525676-ff36-42d8-b6de-408f2c799e37-config\") pod \"service-ca-operator-777779d784-vbcd9\" (UID: \"f4525676-ff36-42d8-b6de-408f2c799e37\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vbcd9" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.145054 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.145133 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5fc8b0b0-d23e-4493-b6f6-6b3036036f39-profile-collector-cert\") pod \"olm-operator-6b444d44fb-8d5rf\" (UID: \"5fc8b0b0-d23e-4493-b6f6-6b3036036f39\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8d5rf" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.145149 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5fc8b0b0-d23e-4493-b6f6-6b3036036f39-srv-cert\") pod \"olm-operator-6b444d44fb-8d5rf\" (UID: \"5fc8b0b0-d23e-4493-b6f6-6b3036036f39\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8d5rf" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.145178 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-td5pg\" (UniqueName: \"kubernetes.io/projected/5fc8b0b0-d23e-4493-b6f6-6b3036036f39-kube-api-access-td5pg\") pod \"olm-operator-6b444d44fb-8d5rf\" (UID: \"5fc8b0b0-d23e-4493-b6f6-6b3036036f39\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8d5rf" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.145204 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-ca-trust-extracted\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.145218 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-installation-pull-secrets\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.145270 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-bound-sa-token\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.145283 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4525676-ff36-42d8-b6de-408f2c799e37-serving-cert\") pod \"service-ca-operator-777779d784-vbcd9\" (UID: \"f4525676-ff36-42d8-b6de-408f2c799e37\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vbcd9" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.145301 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-trusted-ca\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.145334 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-registry-certificates\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.145349 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkzl2\" (UniqueName: \"kubernetes.io/projected/f4525676-ff36-42d8-b6de-408f2c799e37-kube-api-access-vkzl2\") pod \"service-ca-operator-777779d784-vbcd9\" (UID: \"f4525676-ff36-42d8-b6de-408f2c799e37\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vbcd9" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.150673 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r779q" Nov 23 06:47:14 crc kubenswrapper[4559]: E1123 06:47:14.151037 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:14.650701545 +0000 UTC m=+136.672687159 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.162428 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-8wt84"] Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.169938 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zr8cm" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.170405 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-78zq2" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.174846 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-q5d8b"] Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.179668 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-lktfh"] Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.199844 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-ltkhr" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.206666 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tgbrt" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.212097 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xwb9l"] Nov 23 06:47:14 crc kubenswrapper[4559]: W1123 06:47:14.212577 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda91fbd9b_8264_4ac4_b1bd_d62143b7a4d2.slice/crio-6f92c68c0794248f4e5d2b393d73bd3d6fb25870fa1d3891d20041a1481b09cb WatchSource:0}: Error finding container 6f92c68c0794248f4e5d2b393d73bd3d6fb25870fa1d3891d20041a1481b09cb: Status 404 returned error can't find the container with id 6f92c68c0794248f4e5d2b393d73bd3d6fb25870fa1d3891d20041a1481b09cb Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.224712 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2j2z7" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.237999 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-9775t"] Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.246021 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:14 crc kubenswrapper[4559]: E1123 06:47:14.246117 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:14.746098129 +0000 UTC m=+136.768083743 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.246314 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/254348bf-91ce-448f-8bb6-c4f6f72bde3f-csi-data-dir\") pod \"csi-hostpathplugin-2p9ch\" (UID: \"254348bf-91ce-448f-8bb6-c4f6f72bde3f\") " pod="hostpath-provisioner/csi-hostpathplugin-2p9ch" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.246350 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/be99ba37-d853-42ee-bf91-102daae86795-cert\") pod \"ingress-canary-b2l6v\" (UID: \"be99ba37-d853-42ee-bf91-102daae86795\") " pod="openshift-ingress-canary/ingress-canary-b2l6v" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.246386 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-ca-trust-extracted\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.246401 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-installation-pull-secrets\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.246475 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/254348bf-91ce-448f-8bb6-c4f6f72bde3f-socket-dir\") pod \"csi-hostpathplugin-2p9ch\" (UID: \"254348bf-91ce-448f-8bb6-c4f6f72bde3f\") " pod="hostpath-provisioner/csi-hostpathplugin-2p9ch" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.246500 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-bound-sa-token\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.246514 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/624ec98f-bf3b-4651-9d9a-26e6a4377140-config-volume\") pod \"dns-default-mfl2b\" (UID: \"624ec98f-bf3b-4651-9d9a-26e6a4377140\") " pod="openshift-dns/dns-default-mfl2b" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.246528 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4525676-ff36-42d8-b6de-408f2c799e37-serving-cert\") pod \"service-ca-operator-777779d784-vbcd9\" (UID: \"f4525676-ff36-42d8-b6de-408f2c799e37\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vbcd9" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.246592 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-trusted-ca\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.246670 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/254348bf-91ce-448f-8bb6-c4f6f72bde3f-plugins-dir\") pod \"csi-hostpathplugin-2p9ch\" (UID: \"254348bf-91ce-448f-8bb6-c4f6f72bde3f\") " pod="hostpath-provisioner/csi-hostpathplugin-2p9ch" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.246686 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rh7x9\" (UniqueName: \"kubernetes.io/projected/254348bf-91ce-448f-8bb6-c4f6f72bde3f-kube-api-access-rh7x9\") pod \"csi-hostpathplugin-2p9ch\" (UID: \"254348bf-91ce-448f-8bb6-c4f6f72bde3f\") " pod="hostpath-provisioner/csi-hostpathplugin-2p9ch" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.247389 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-ca-trust-extracted\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.248331 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-registry-certificates\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.248760 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkzl2\" (UniqueName: \"kubernetes.io/projected/f4525676-ff36-42d8-b6de-408f2c799e37-kube-api-access-vkzl2\") pod \"service-ca-operator-777779d784-vbcd9\" (UID: \"f4525676-ff36-42d8-b6de-408f2c799e37\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vbcd9" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.251123 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4525676-ff36-42d8-b6de-408f2c799e37-serving-cert\") pod \"service-ca-operator-777779d784-vbcd9\" (UID: \"f4525676-ff36-42d8-b6de-408f2c799e37\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vbcd9" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.251173 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-registry-certificates\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.251282 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-trusted-ca\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.251303 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rg7c6\" (UniqueName: \"kubernetes.io/projected/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-kube-api-access-rg7c6\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.251600 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-registry-tls\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.251698 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4525676-ff36-42d8-b6de-408f2c799e37-config\") pod \"service-ca-operator-777779d784-vbcd9\" (UID: \"f4525676-ff36-42d8-b6de-408f2c799e37\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vbcd9" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.251801 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:14 crc kubenswrapper[4559]: E1123 06:47:14.252112 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:14.752100396 +0000 UTC m=+136.774086010 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.252188 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4525676-ff36-42d8-b6de-408f2c799e37-config\") pod \"service-ca-operator-777779d784-vbcd9\" (UID: \"f4525676-ff36-42d8-b6de-408f2c799e37\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vbcd9" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.252324 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/254348bf-91ce-448f-8bb6-c4f6f72bde3f-registration-dir\") pod \"csi-hostpathplugin-2p9ch\" (UID: \"254348bf-91ce-448f-8bb6-c4f6f72bde3f\") " pod="hostpath-provisioner/csi-hostpathplugin-2p9ch" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.252389 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-installation-pull-secrets\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.252988 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/624ec98f-bf3b-4651-9d9a-26e6a4377140-metrics-tls\") pod \"dns-default-mfl2b\" (UID: \"624ec98f-bf3b-4651-9d9a-26e6a4377140\") " pod="openshift-dns/dns-default-mfl2b" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.253136 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-449cc\" (UniqueName: \"kubernetes.io/projected/be99ba37-d853-42ee-bf91-102daae86795-kube-api-access-449cc\") pod \"ingress-canary-b2l6v\" (UID: \"be99ba37-d853-42ee-bf91-102daae86795\") " pod="openshift-ingress-canary/ingress-canary-b2l6v" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.253580 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5fc8b0b0-d23e-4493-b6f6-6b3036036f39-profile-collector-cert\") pod \"olm-operator-6b444d44fb-8d5rf\" (UID: \"5fc8b0b0-d23e-4493-b6f6-6b3036036f39\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8d5rf" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.253601 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/254348bf-91ce-448f-8bb6-c4f6f72bde3f-mountpoint-dir\") pod \"csi-hostpathplugin-2p9ch\" (UID: \"254348bf-91ce-448f-8bb6-c4f6f72bde3f\") " pod="hostpath-provisioner/csi-hostpathplugin-2p9ch" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.253654 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5fc8b0b0-d23e-4493-b6f6-6b3036036f39-srv-cert\") pod \"olm-operator-6b444d44fb-8d5rf\" (UID: \"5fc8b0b0-d23e-4493-b6f6-6b3036036f39\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8d5rf" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.253671 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsjkr\" (UniqueName: \"kubernetes.io/projected/624ec98f-bf3b-4651-9d9a-26e6a4377140-kube-api-access-zsjkr\") pod \"dns-default-mfl2b\" (UID: \"624ec98f-bf3b-4651-9d9a-26e6a4377140\") " pod="openshift-dns/dns-default-mfl2b" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.253703 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-td5pg\" (UniqueName: \"kubernetes.io/projected/5fc8b0b0-d23e-4493-b6f6-6b3036036f39-kube-api-access-td5pg\") pod \"olm-operator-6b444d44fb-8d5rf\" (UID: \"5fc8b0b0-d23e-4493-b6f6-6b3036036f39\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8d5rf" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.256309 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-registry-tls\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:14 crc kubenswrapper[4559]: W1123 06:47:14.257287 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50e7a859_2c04_4eff_889d_633ced116f34.slice/crio-853b3dfcb13bc9f311931394550e1c84a9423b32ce99da0dfe2f1ae22904a7fa WatchSource:0}: Error finding container 853b3dfcb13bc9f311931394550e1c84a9423b32ce99da0dfe2f1ae22904a7fa: Status 404 returned error can't find the container with id 853b3dfcb13bc9f311931394550e1c84a9423b32ce99da0dfe2f1ae22904a7fa Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.257623 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5fc8b0b0-d23e-4493-b6f6-6b3036036f39-srv-cert\") pod \"olm-operator-6b444d44fb-8d5rf\" (UID: \"5fc8b0b0-d23e-4493-b6f6-6b3036036f39\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8d5rf" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.262014 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5fc8b0b0-d23e-4493-b6f6-6b3036036f39-profile-collector-cert\") pod \"olm-operator-6b444d44fb-8d5rf\" (UID: \"5fc8b0b0-d23e-4493-b6f6-6b3036036f39\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8d5rf" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.287298 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-bound-sa-token\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.306456 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkzl2\" (UniqueName: \"kubernetes.io/projected/f4525676-ff36-42d8-b6de-408f2c799e37-kube-api-access-vkzl2\") pod \"service-ca-operator-777779d784-vbcd9\" (UID: \"f4525676-ff36-42d8-b6de-408f2c799e37\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vbcd9" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.334990 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rg7c6\" (UniqueName: \"kubernetes.io/projected/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-kube-api-access-rg7c6\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.346705 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-td5pg\" (UniqueName: \"kubernetes.io/projected/5fc8b0b0-d23e-4493-b6f6-6b3036036f39-kube-api-access-td5pg\") pod \"olm-operator-6b444d44fb-8d5rf\" (UID: \"5fc8b0b0-d23e-4493-b6f6-6b3036036f39\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8d5rf" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.356224 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:14 crc kubenswrapper[4559]: E1123 06:47:14.356398 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:14.856379226 +0000 UTC m=+136.878364841 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.356931 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.356964 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/254348bf-91ce-448f-8bb6-c4f6f72bde3f-registration-dir\") pod \"csi-hostpathplugin-2p9ch\" (UID: \"254348bf-91ce-448f-8bb6-c4f6f72bde3f\") " pod="hostpath-provisioner/csi-hostpathplugin-2p9ch" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.356986 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/624ec98f-bf3b-4651-9d9a-26e6a4377140-metrics-tls\") pod \"dns-default-mfl2b\" (UID: \"624ec98f-bf3b-4651-9d9a-26e6a4377140\") " pod="openshift-dns/dns-default-mfl2b" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.357004 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-449cc\" (UniqueName: \"kubernetes.io/projected/be99ba37-d853-42ee-bf91-102daae86795-kube-api-access-449cc\") pod \"ingress-canary-b2l6v\" (UID: \"be99ba37-d853-42ee-bf91-102daae86795\") " pod="openshift-ingress-canary/ingress-canary-b2l6v" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.357035 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/254348bf-91ce-448f-8bb6-c4f6f72bde3f-mountpoint-dir\") pod \"csi-hostpathplugin-2p9ch\" (UID: \"254348bf-91ce-448f-8bb6-c4f6f72bde3f\") " pod="hostpath-provisioner/csi-hostpathplugin-2p9ch" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.357058 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsjkr\" (UniqueName: \"kubernetes.io/projected/624ec98f-bf3b-4651-9d9a-26e6a4377140-kube-api-access-zsjkr\") pod \"dns-default-mfl2b\" (UID: \"624ec98f-bf3b-4651-9d9a-26e6a4377140\") " pod="openshift-dns/dns-default-mfl2b" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.357071 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/254348bf-91ce-448f-8bb6-c4f6f72bde3f-csi-data-dir\") pod \"csi-hostpathplugin-2p9ch\" (UID: \"254348bf-91ce-448f-8bb6-c4f6f72bde3f\") " pod="hostpath-provisioner/csi-hostpathplugin-2p9ch" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.357088 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/be99ba37-d853-42ee-bf91-102daae86795-cert\") pod \"ingress-canary-b2l6v\" (UID: \"be99ba37-d853-42ee-bf91-102daae86795\") " pod="openshift-ingress-canary/ingress-canary-b2l6v" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.357117 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/254348bf-91ce-448f-8bb6-c4f6f72bde3f-socket-dir\") pod \"csi-hostpathplugin-2p9ch\" (UID: \"254348bf-91ce-448f-8bb6-c4f6f72bde3f\") " pod="hostpath-provisioner/csi-hostpathplugin-2p9ch" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.357132 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/624ec98f-bf3b-4651-9d9a-26e6a4377140-config-volume\") pod \"dns-default-mfl2b\" (UID: \"624ec98f-bf3b-4651-9d9a-26e6a4377140\") " pod="openshift-dns/dns-default-mfl2b" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.357152 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/254348bf-91ce-448f-8bb6-c4f6f72bde3f-plugins-dir\") pod \"csi-hostpathplugin-2p9ch\" (UID: \"254348bf-91ce-448f-8bb6-c4f6f72bde3f\") " pod="hostpath-provisioner/csi-hostpathplugin-2p9ch" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.357164 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rh7x9\" (UniqueName: \"kubernetes.io/projected/254348bf-91ce-448f-8bb6-c4f6f72bde3f-kube-api-access-rh7x9\") pod \"csi-hostpathplugin-2p9ch\" (UID: \"254348bf-91ce-448f-8bb6-c4f6f72bde3f\") " pod="hostpath-provisioner/csi-hostpathplugin-2p9ch" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.357161 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/254348bf-91ce-448f-8bb6-c4f6f72bde3f-registration-dir\") pod \"csi-hostpathplugin-2p9ch\" (UID: \"254348bf-91ce-448f-8bb6-c4f6f72bde3f\") " pod="hostpath-provisioner/csi-hostpathplugin-2p9ch" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.357241 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/254348bf-91ce-448f-8bb6-c4f6f72bde3f-csi-data-dir\") pod \"csi-hostpathplugin-2p9ch\" (UID: \"254348bf-91ce-448f-8bb6-c4f6f72bde3f\") " pod="hostpath-provisioner/csi-hostpathplugin-2p9ch" Nov 23 06:47:14 crc kubenswrapper[4559]: E1123 06:47:14.357334 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:14.857319692 +0000 UTC m=+136.879305306 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.357394 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/254348bf-91ce-448f-8bb6-c4f6f72bde3f-mountpoint-dir\") pod \"csi-hostpathplugin-2p9ch\" (UID: \"254348bf-91ce-448f-8bb6-c4f6f72bde3f\") " pod="hostpath-provisioner/csi-hostpathplugin-2p9ch" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.357572 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/254348bf-91ce-448f-8bb6-c4f6f72bde3f-plugins-dir\") pod \"csi-hostpathplugin-2p9ch\" (UID: \"254348bf-91ce-448f-8bb6-c4f6f72bde3f\") " pod="hostpath-provisioner/csi-hostpathplugin-2p9ch" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.357621 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/254348bf-91ce-448f-8bb6-c4f6f72bde3f-socket-dir\") pod \"csi-hostpathplugin-2p9ch\" (UID: \"254348bf-91ce-448f-8bb6-c4f6f72bde3f\") " pod="hostpath-provisioner/csi-hostpathplugin-2p9ch" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.357985 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/624ec98f-bf3b-4651-9d9a-26e6a4377140-config-volume\") pod \"dns-default-mfl2b\" (UID: \"624ec98f-bf3b-4651-9d9a-26e6a4377140\") " pod="openshift-dns/dns-default-mfl2b" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.359729 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/624ec98f-bf3b-4651-9d9a-26e6a4377140-metrics-tls\") pod \"dns-default-mfl2b\" (UID: \"624ec98f-bf3b-4651-9d9a-26e6a4377140\") " pod="openshift-dns/dns-default-mfl2b" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.361004 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/be99ba37-d853-42ee-bf91-102daae86795-cert\") pod \"ingress-canary-b2l6v\" (UID: \"be99ba37-d853-42ee-bf91-102daae86795\") " pod="openshift-ingress-canary/ingress-canary-b2l6v" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.412407 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-449cc\" (UniqueName: \"kubernetes.io/projected/be99ba37-d853-42ee-bf91-102daae86795-kube-api-access-449cc\") pod \"ingress-canary-b2l6v\" (UID: \"be99ba37-d853-42ee-bf91-102daae86795\") " pod="openshift-ingress-canary/ingress-canary-b2l6v" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.445096 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsjkr\" (UniqueName: \"kubernetes.io/projected/624ec98f-bf3b-4651-9d9a-26e6a4377140-kube-api-access-zsjkr\") pod \"dns-default-mfl2b\" (UID: \"624ec98f-bf3b-4651-9d9a-26e6a4377140\") " pod="openshift-dns/dns-default-mfl2b" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.452842 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rh7x9\" (UniqueName: \"kubernetes.io/projected/254348bf-91ce-448f-8bb6-c4f6f72bde3f-kube-api-access-rh7x9\") pod \"csi-hostpathplugin-2p9ch\" (UID: \"254348bf-91ce-448f-8bb6-c4f6f72bde3f\") " pod="hostpath-provisioner/csi-hostpathplugin-2p9ch" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.454869 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-dnz8k"] Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.460177 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:14 crc kubenswrapper[4559]: E1123 06:47:14.460453 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:14.960439852 +0000 UTC m=+136.982425466 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.478074 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398005-nnldq"] Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.485730 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mkb95"] Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.489281 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vbcd9" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.533095 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8d5rf" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.560079 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zr8cm"] Nov 23 06:47:14 crc kubenswrapper[4559]: W1123 06:47:14.560295 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ac1ccfb_fbd3_4de6_83bd_ff1ddcd531eb.slice/crio-67be934335d4ec36a6eab4685d6e55efa96e8469ac039bd2a901390f9513299a WatchSource:0}: Error finding container 67be934335d4ec36a6eab4685d6e55efa96e8469ac039bd2a901390f9513299a: Status 404 returned error can't find the container with id 67be934335d4ec36a6eab4685d6e55efa96e8469ac039bd2a901390f9513299a Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.561603 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:14 crc kubenswrapper[4559]: E1123 06:47:14.561847 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:15.06183714 +0000 UTC m=+137.083822755 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.575875 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-2p9ch" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.581630 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-mfl2b" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.584170 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-b2l6v" Nov 23 06:47:14 crc kubenswrapper[4559]: W1123 06:47:14.593955 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2879a16a_7ab6_413a_a8a3_ecfa7ffd4ac7.slice/crio-61f3e4059c92971913f7fbebeaf12e6c260763b93f60e042d4027c1b9b66706d WatchSource:0}: Error finding container 61f3e4059c92971913f7fbebeaf12e6c260763b93f60e042d4027c1b9b66706d: Status 404 returned error can't find the container with id 61f3e4059c92971913f7fbebeaf12e6c260763b93f60e042d4027c1b9b66706d Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.659025 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-t2rfc"] Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.674229 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:14 crc kubenswrapper[4559]: E1123 06:47:14.674529 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:15.17451561 +0000 UTC m=+137.196501224 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.681754 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tgbrt"] Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.697298 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qt4fl"] Nov 23 06:47:14 crc kubenswrapper[4559]: W1123 06:47:14.722011 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a922bdd_f7be_41d2_8626_9352ec57d9ce.slice/crio-fee5a88bed9770e02ed8b33343744955549d6b55a3bd6dac28ed9a3336642a40 WatchSource:0}: Error finding container fee5a88bed9770e02ed8b33343744955549d6b55a3bd6dac28ed9a3336642a40: Status 404 returned error can't find the container with id fee5a88bed9770e02ed8b33343744955549d6b55a3bd6dac28ed9a3336642a40 Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.775852 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:14 crc kubenswrapper[4559]: E1123 06:47:14.776291 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:15.276281108 +0000 UTC m=+137.298266721 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:14 crc kubenswrapper[4559]: W1123 06:47:14.777337 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09d28a3a_5b25_4e31_90d6_17d9319cd11c.slice/crio-ab4872c279f316d22f44bfa5e54043eee7b74e09a268baca0534faa406395e55 WatchSource:0}: Error finding container ab4872c279f316d22f44bfa5e54043eee7b74e09a268baca0534faa406395e55: Status 404 returned error can't find the container with id ab4872c279f316d22f44bfa5e54043eee7b74e09a268baca0534faa406395e55 Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.780749 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-ztnsp" event={"ID":"18f99d77-1c2d-412b-93ae-1d1fc52f24ab","Type":"ContainerStarted","Data":"e6006e8a4df21bafcfc14739937a750bddb5536b03f6bcb88bc8d80ee7ee2b2c"} Nov 23 06:47:14 crc kubenswrapper[4559]: W1123 06:47:14.781003 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80d890b8_c797_4392_a48a_7894059147d4.slice/crio-f8e84e797a5cff7745974e41b07434d4fe880c30f146a9e2838f782c05cfb22a WatchSource:0}: Error finding container f8e84e797a5cff7745974e41b07434d4fe880c30f146a9e2838f782c05cfb22a: Status 404 returned error can't find the container with id f8e84e797a5cff7745974e41b07434d4fe880c30f146a9e2838f782c05cfb22a Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.792264 4559 generic.go:334] "Generic (PLEG): container finished" podID="e814799e-e374-4ad8-96a2-46e4d5defc2f" containerID="a5a0d8973657d9a4bfbcd8604db0c8ca0db054b1d8fc34df60e009a8bb4d101e" exitCode=0 Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.792334 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" event={"ID":"e814799e-e374-4ad8-96a2-46e4d5defc2f","Type":"ContainerDied","Data":"a5a0d8973657d9a4bfbcd8604db0c8ca0db054b1d8fc34df60e009a8bb4d101e"} Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.792359 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" event={"ID":"e814799e-e374-4ad8-96a2-46e4d5defc2f","Type":"ContainerStarted","Data":"8dddf16b1e7b7e259df778cf1b48c434c43cb8672dcad0e7dcf6e855c0c1214a"} Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.813513 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b" event={"ID":"08b0652b-f30f-4c2c-bd1a-75b80a66b144","Type":"ContainerStarted","Data":"742c23d558702aa7bdf98ee69c5e2808ee4470e0935d2b09dbd82585cd894eeb"} Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.813545 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.813572 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b" event={"ID":"08b0652b-f30f-4c2c-bd1a-75b80a66b144","Type":"ContainerStarted","Data":"e65dfed27469c0d817ec33a4ee61e8bb93ff8846276827e3968f8dc3cfc014c9"} Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.825091 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.827508 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" event={"ID":"e7b5892a-8920-4cef-94a1-3736328e9a39","Type":"ContainerStarted","Data":"ce69456f4d4198e8b9edc01bd976a0de7270cf5baf44e52b38599f01a9ab6c82"} Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.828294 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.845023 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.859962 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8xfrw" event={"ID":"a130a936-3db2-4455-bb63-fcea37eda82c","Type":"ContainerStarted","Data":"6ed60a9b745884d6ba66e70915f23d57d00485a040f8abdde9ac24fb315804aa"} Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.863355 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-q5d8b" event={"ID":"7f90213c-484b-4f9d-b064-363887589530","Type":"ContainerStarted","Data":"7f5c9ad2ac60ce9ff0b85fe57ae8e3bc92b02ff8e63fc772bc58ebdb62b3d135"} Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.876625 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:14 crc kubenswrapper[4559]: E1123 06:47:14.877100 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:15.377087544 +0000 UTC m=+137.399073158 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.877213 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-r5ts6" event={"ID":"534e137b-1c8d-45fb-9c1b-6804913d309b","Type":"ContainerStarted","Data":"474a804b27f6b4b95c796f5b32a7a67419f708adfb40b8b4739183d0ce9df8cc"} Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.877236 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-r5ts6" event={"ID":"534e137b-1c8d-45fb-9c1b-6804913d309b","Type":"ContainerStarted","Data":"72d20e034c518fb5213fbccb1e48ed36b5df5d250f4943f74f1390eef8af6220"} Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.877738 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-r5ts6" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.890563 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-jgl4n" event={"ID":"3d1342fa-3bb4-4924-a36b-e370aec66cd3","Type":"ContainerStarted","Data":"70813a6129c83720c173e987d3c569d4fd87143b60c0d1e2ee1c9ead754d7b7a"} Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.890588 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-jgl4n" event={"ID":"3d1342fa-3bb4-4924-a36b-e370aec66cd3","Type":"ContainerStarted","Data":"72e1da1bfcad1c11124429b3e0c99919e768a717e98b56ffd26e68db7adbb56a"} Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.921834 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-r5ts6" Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.932758 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-ltkhr"] Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.955157 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ggt5f" event={"ID":"35b44360-64aa-45b1-9141-15d5f69b384a","Type":"ContainerStarted","Data":"fe3e14d63e82975eb1b6146854f7b91c7d25cb18fcf751e01ea2bb0274d42940"} Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.960316 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-dlj2h" event={"ID":"16e9fcf3-adc7-4af9-a625-125ff496cfcf","Type":"ContainerStarted","Data":"b416dd75f2688c75a184292df271196b89e28090b1dd068df5498ed9dc24e25b"} Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.960360 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-dlj2h" event={"ID":"16e9fcf3-adc7-4af9-a625-125ff496cfcf","Type":"ContainerStarted","Data":"4646bbe0ef780288484acdf5ebe576523bd7f0deb1b75d6733042590686b35e9"} Nov 23 06:47:14 crc kubenswrapper[4559]: I1123 06:47:14.977808 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:14 crc kubenswrapper[4559]: E1123 06:47:14.979920 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:15.479908446 +0000 UTC m=+137.501894059 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.006197 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r779q"] Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.009819 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8llfp" event={"ID":"b3e2ea56-3525-4e78-a7c8-289a9a2406e6","Type":"ContainerStarted","Data":"d7bd5ebd161e1f2a662f13484a934ceba6e8f8823b8d3958b250e704902831b8"} Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.009856 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8llfp" event={"ID":"b3e2ea56-3525-4e78-a7c8-289a9a2406e6","Type":"ContainerStarted","Data":"baef0316cbcd6a289f7121303a4af428ebfd41b3980e331b71de04a2e2a99289"} Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.016899 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2j2z7"] Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.036163 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-78zq2"] Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.036959 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-rqftc" event={"ID":"41ae564b-da35-45eb-9a27-fbf23bacaf8f","Type":"ContainerStarted","Data":"cbf2a6e73edb663686a2506b4cb671c3645f2fe1bc687ba4639fde99e1303f0c"} Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.037014 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-rqftc" event={"ID":"41ae564b-da35-45eb-9a27-fbf23bacaf8f","Type":"ContainerStarted","Data":"1a4703895e0a1f246b78c82e5cd28ad65d1c93bff2cd865b4313043f231a0db7"} Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.037880 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-rqftc" Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.042797 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-nnldq" event={"ID":"d1d0d430-031f-448e-9763-69deb17abc4a","Type":"ContainerStarted","Data":"e947b16875ec3b22595f0769f7b02bd01da5259a48b5d893ca321c2a3df3db1c"} Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.061774 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-9gwbh" event={"ID":"a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba","Type":"ContainerStarted","Data":"912ba031d905f731b0a3a7f1c4dbf52f8ded946420dc47663fea46de85972c07"} Nov 23 06:47:15 crc kubenswrapper[4559]: W1123 06:47:15.074805 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod04abe9e5_6ffe_429d_9274_73d133fd0d52.slice/crio-9435f802d3ffbe8940df2e6f005384631945bc557b5530a36c070135f1ad9c84 WatchSource:0}: Error finding container 9435f802d3ffbe8940df2e6f005384631945bc557b5530a36c070135f1ad9c84: Status 404 returned error can't find the container with id 9435f802d3ffbe8940df2e6f005384631945bc557b5530a36c070135f1ad9c84 Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.080872 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:15 crc kubenswrapper[4559]: E1123 06:47:15.081134 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:15.581120403 +0000 UTC m=+137.603106017 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.081375 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.082260 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fjv2r" event={"ID":"e417e794-f6b2-4607-b2ce-3a737e8770b4","Type":"ContainerStarted","Data":"0e0f701368bcb381d71ad9a56ff021e108914328946fbc315a9e12fcb4010f1b"} Nov 23 06:47:15 crc kubenswrapper[4559]: E1123 06:47:15.082906 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:15.582897056 +0000 UTC m=+137.604882670 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.089283 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" event={"ID":"994432aa-8794-4818-94bd-ce28d97c7163","Type":"ContainerStarted","Data":"8962345c29b078925ab6d68c9a56651f2587982e22f1156c34f40a8ad0aabf5a"} Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.090150 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xwb9l" event={"ID":"1b7acf5d-cae8-429a-8661-ec3abcf056c0","Type":"ContainerStarted","Data":"276e3364169cdae9d9f30800a190dc92cb92871197f1bb2c5009773a4ff4decc"} Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.096902 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-p9rps" event={"ID":"4316762b-b889-41c0-aa31-f874bc0fa3fe","Type":"ContainerStarted","Data":"a9e7974372b448d34b591b6666dad86e825bbee6371db2cc78ebb61fc4fb5a8a"} Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.096945 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-p9rps" event={"ID":"4316762b-b889-41c0-aa31-f874bc0fa3fe","Type":"ContainerStarted","Data":"a58c30ab9a2f436fbb7e56b2da572ec27b2c40aa7afd8d94dcb5bb443f3ee06c"} Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.097046 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-p9rps" Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.104195 4559 patch_prober.go:28] interesting pod/downloads-7954f5f757-p9rps container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.104234 4559 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-p9rps" podUID="4316762b-b889-41c0-aa31-f874bc0fa3fe" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.125847 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mkb95" event={"ID":"0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb","Type":"ContainerStarted","Data":"67be934335d4ec36a6eab4685d6e55efa96e8469ac039bd2a901390f9513299a"} Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.178116 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lktfh" event={"ID":"50e7a859-2c04-4eff-889d-633ced116f34","Type":"ContainerStarted","Data":"853b3dfcb13bc9f311931394550e1c84a9423b32ce99da0dfe2f1ae22904a7fa"} Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.190197 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:15 crc kubenswrapper[4559]: E1123 06:47:15.191098 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:15.691084768 +0000 UTC m=+137.713070382 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.195912 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-9m2sk" event={"ID":"9e9514f9-f14f-4409-8fef-b89a37de70a6","Type":"ContainerStarted","Data":"a78fe92961ddc78d9c7ee657ca51ea444c30d146b657f0c8cbfa21cd11b6efcb"} Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.195965 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-9m2sk" event={"ID":"9e9514f9-f14f-4409-8fef-b89a37de70a6","Type":"ContainerStarted","Data":"715425ac79955143a59d44b34a66675c0774df61787cb9f4f80ab7d70b17b176"} Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.206252 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7z49" podStartSLOduration=116.20623708 podStartE2EDuration="1m56.20623708s" podCreationTimestamp="2025-11-23 06:45:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:15.206028634 +0000 UTC m=+137.228014249" watchObservedRunningTime="2025-11-23 06:47:15.20623708 +0000 UTC m=+137.228222694" Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.220252 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8d5rf"] Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.227395 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dnz8k" event={"ID":"2879a16a-7ab6-413a-a8a3-ecfa7ffd4ac7","Type":"ContainerStarted","Data":"61f3e4059c92971913f7fbebeaf12e6c260763b93f60e042d4027c1b9b66706d"} Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.229101 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-8wt84" event={"ID":"a91fbd9b-8264-4ac4-b1bd-d62143b7a4d2","Type":"ContainerStarted","Data":"6f92c68c0794248f4e5d2b393d73bd3d6fb25870fa1d3891d20041a1481b09cb"} Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.231623 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j665r" event={"ID":"58f094f8-8c16-4929-ae94-e52241c5b7d1","Type":"ContainerStarted","Data":"178c760a704f3f7c6f406f79eb7da0aed8f6d343ebb3267711043c5a8e105e70"} Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.231665 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j665r" event={"ID":"58f094f8-8c16-4929-ae94-e52241c5b7d1","Type":"ContainerStarted","Data":"9883ac7de2fa8b610bf6b30e6c28b1656b6f88e662f4f9c4d18c6303bda1707b"} Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.261836 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-xvd4d" event={"ID":"6796a9bd-a428-4b0a-8723-d8a6a252de84","Type":"ContainerStarted","Data":"e813441ff551059a86acca42beb61e7154ff1ea903a964f49529e56b8afc01fc"} Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.291287 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:15 crc kubenswrapper[4559]: E1123 06:47:15.292356 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:15.792345107 +0000 UTC m=+137.814330721 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.310436 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9775t" event={"ID":"266931fd-4205-4e31-9ded-9882a2a41921","Type":"ContainerStarted","Data":"cced9721f3a4950b5b0a0b4a6dd45ea168bc3d860e5034d704b7ee68702924d5"} Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.316297 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q79mk" event={"ID":"3a883895-9775-4c18-962e-be4df71043c8","Type":"ContainerStarted","Data":"dbade96c604581e6d517c7b0ec082ce7fb58fbfd4c8975e0b00e540b5ad35d74"} Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.316322 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q79mk" event={"ID":"3a883895-9775-4c18-962e-be4df71043c8","Type":"ContainerStarted","Data":"bf95d3d7e9583d0154f6284e63286326f42ec03359abf6eadcd04418964a8f14"} Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.391967 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:15 crc kubenswrapper[4559]: E1123 06:47:15.392536 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:15.892522669 +0000 UTC m=+137.914508283 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.502020 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:15 crc kubenswrapper[4559]: E1123 06:47:15.503611 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:16.003600048 +0000 UTC m=+138.025585662 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.608065 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:15 crc kubenswrapper[4559]: E1123 06:47:15.608823 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:16.10880661 +0000 UTC m=+138.130792224 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.680443 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-mfl2b"] Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.682589 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2p9ch"] Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.713580 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:15 crc kubenswrapper[4559]: E1123 06:47:15.713832 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:16.213822108 +0000 UTC m=+138.235807722 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.765588 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vbcd9"] Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.814996 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:15 crc kubenswrapper[4559]: E1123 06:47:15.815215 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:16.315195893 +0000 UTC m=+138.337181507 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.815411 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:15 crc kubenswrapper[4559]: E1123 06:47:15.815802 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:16.315790432 +0000 UTC m=+138.337776046 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.831141 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-rqftc" Nov 23 06:47:15 crc kubenswrapper[4559]: W1123 06:47:15.856050 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4525676_ff36_42d8_b6de_408f2c799e37.slice/crio-373ff90fa9608bddbcab1925af502079a0055959a6c02c1520d26cb6c1eba726 WatchSource:0}: Error finding container 373ff90fa9608bddbcab1925af502079a0055959a6c02c1520d26cb6c1eba726: Status 404 returned error can't find the container with id 373ff90fa9608bddbcab1925af502079a0055959a6c02c1520d26cb6c1eba726 Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.860037 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-b2l6v"] Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.919867 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:15 crc kubenswrapper[4559]: E1123 06:47:15.920276 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:16.420260324 +0000 UTC m=+138.442245938 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.928333 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-jgl4n" podStartSLOduration=4.928319597 podStartE2EDuration="4.928319597s" podCreationTimestamp="2025-11-23 06:47:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:15.926808669 +0000 UTC m=+137.948794282" watchObservedRunningTime="2025-11-23 06:47:15.928319597 +0000 UTC m=+137.950305212" Nov 23 06:47:15 crc kubenswrapper[4559]: I1123 06:47:15.965534 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-xvd4d" podStartSLOduration=115.965517852 podStartE2EDuration="1m55.965517852s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:15.965517841 +0000 UTC m=+137.987503455" watchObservedRunningTime="2025-11-23 06:47:15.965517852 +0000 UTC m=+137.987503465" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.022698 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:16 crc kubenswrapper[4559]: E1123 06:47:16.022967 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:16.522944807 +0000 UTC m=+138.544930421 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.056359 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-r5ts6" podStartSLOduration=116.056344808 podStartE2EDuration="1m56.056344808s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:16.019742056 +0000 UTC m=+138.041727670" watchObservedRunningTime="2025-11-23 06:47:16.056344808 +0000 UTC m=+138.078330422" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.084557 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-rqftc" podStartSLOduration=116.084543283 podStartE2EDuration="1m56.084543283s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:16.082430291 +0000 UTC m=+138.104415905" watchObservedRunningTime="2025-11-23 06:47:16.084543283 +0000 UTC m=+138.106528897" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.123362 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:16 crc kubenswrapper[4559]: E1123 06:47:16.123724 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:16.623710576 +0000 UTC m=+138.645696190 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.130590 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q79mk" podStartSLOduration=117.130575472 podStartE2EDuration="1m57.130575472s" podCreationTimestamp="2025-11-23 06:45:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:16.124023681 +0000 UTC m=+138.146009295" watchObservedRunningTime="2025-11-23 06:47:16.130575472 +0000 UTC m=+138.152561085" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.184238 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j665r" podStartSLOduration=116.184220074 podStartE2EDuration="1m56.184220074s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:16.175107571 +0000 UTC m=+138.197093185" watchObservedRunningTime="2025-11-23 06:47:16.184220074 +0000 UTC m=+138.206205708" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.225444 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:16 crc kubenswrapper[4559]: E1123 06:47:16.225869 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:16.725855945 +0000 UTC m=+138.747841559 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.227675 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ggt5f" podStartSLOduration=116.227663428 podStartE2EDuration="1m56.227663428s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:16.226154081 +0000 UTC m=+138.248139695" watchObservedRunningTime="2025-11-23 06:47:16.227663428 +0000 UTC m=+138.249649041" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.336109 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:16 crc kubenswrapper[4559]: E1123 06:47:16.336507 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:16.836491445 +0000 UTC m=+138.858477059 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.337295 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-q5d8b" podStartSLOduration=117.337284801 podStartE2EDuration="1m57.337284801s" podCreationTimestamp="2025-11-23 06:45:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:16.268740957 +0000 UTC m=+138.290726571" watchObservedRunningTime="2025-11-23 06:47:16.337284801 +0000 UTC m=+138.359270416" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.380170 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-ztnsp" event={"ID":"18f99d77-1c2d-412b-93ae-1d1fc52f24ab","Type":"ContainerStarted","Data":"81e2a27cf50392c93fdd6fc79e9baaa43c8e2b6f16dcbd0da7718c865077d441"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.380460 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-ztnsp" event={"ID":"18f99d77-1c2d-412b-93ae-1d1fc52f24ab","Type":"ContainerStarted","Data":"aca345f08ec477465321a8c2398fa99ce653b149ca3ff4315969d931f61b0f92"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.380341 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b" podStartSLOduration=116.38032577 podStartE2EDuration="1m56.38032577s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:16.338188007 +0000 UTC m=+138.360173621" watchObservedRunningTime="2025-11-23 06:47:16.38032577 +0000 UTC m=+138.402311384" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.397156 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r779q" event={"ID":"04abe9e5-6ffe-429d-9274-73d133fd0d52","Type":"ContainerStarted","Data":"ed840eb91406022fde2ac3d442bd2245723e7f800d20344090d5ab1fc391c34f"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.397198 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r779q" event={"ID":"04abe9e5-6ffe-429d-9274-73d133fd0d52","Type":"ContainerStarted","Data":"9435f802d3ffbe8940df2e6f005384631945bc557b5530a36c070135f1ad9c84"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.437321 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:16 crc kubenswrapper[4559]: E1123 06:47:16.437635 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:16.937624181 +0000 UTC m=+138.959609796 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.445156 4559 generic.go:334] "Generic (PLEG): container finished" podID="e417e794-f6b2-4607-b2ce-3a737e8770b4" containerID="39920a0ecbf31cb53b03a360c14ab98037accede808bebf5eb34fd87432e0dcd" exitCode=0 Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.445273 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fjv2r" event={"ID":"e417e794-f6b2-4607-b2ce-3a737e8770b4","Type":"ContainerStarted","Data":"13600b92215f04f5a9612c59d3e16ca500929db249617a3463e645bbda46b34c"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.445318 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fjv2r" event={"ID":"e417e794-f6b2-4607-b2ce-3a737e8770b4","Type":"ContainerDied","Data":"39920a0ecbf31cb53b03a360c14ab98037accede808bebf5eb34fd87432e0dcd"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.445689 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fjv2r" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.450246 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-p9rps" podStartSLOduration=116.450233144 podStartE2EDuration="1m56.450233144s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:16.396906966 +0000 UTC m=+138.418892580" watchObservedRunningTime="2025-11-23 06:47:16.450233144 +0000 UTC m=+138.472218759" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.451300 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8llfp" podStartSLOduration=116.451292566 podStartE2EDuration="1m56.451292566s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:16.445518833 +0000 UTC m=+138.467504447" watchObservedRunningTime="2025-11-23 06:47:16.451292566 +0000 UTC m=+138.473278180" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.491981 4559 generic.go:334] "Generic (PLEG): container finished" podID="994432aa-8794-4818-94bd-ce28d97c7163" containerID="d50e063f8c074f3c16f791988573e1844ef39cf7a71377ef59966bc36fbd0d7d" exitCode=0 Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.492050 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" event={"ID":"994432aa-8794-4818-94bd-ce28d97c7163","Type":"ContainerDied","Data":"d50e063f8c074f3c16f791988573e1844ef39cf7a71377ef59966bc36fbd0d7d"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.529914 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dnz8k" event={"ID":"2879a16a-7ab6-413a-a8a3-ecfa7ffd4ac7","Type":"ContainerStarted","Data":"5e0f45d73df1770e2e1d618794483c6e6c3e196ba489f233e141f56052bf1d6f"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.529967 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dnz8k" event={"ID":"2879a16a-7ab6-413a-a8a3-ecfa7ffd4ac7","Type":"ContainerStarted","Data":"4cb75c23434014875e77cad7317c2eacc876d0e04611a3d37bf03ddc7ca9abfe"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.541022 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:16 crc kubenswrapper[4559]: E1123 06:47:16.541160 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:17.041120716 +0000 UTC m=+139.063106329 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.541328 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:16 crc kubenswrapper[4559]: E1123 06:47:16.541675 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:17.041662484 +0000 UTC m=+139.063648089 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.544898 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zr8cm" event={"ID":"3a922bdd-f7be-41d2-8626-9352ec57d9ce","Type":"ContainerStarted","Data":"6d281c956d4fb39415e675b1d3f42222560359e86c96aecfe88ce07028744929"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.544934 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zr8cm" event={"ID":"3a922bdd-f7be-41d2-8626-9352ec57d9ce","Type":"ContainerStarted","Data":"fee5a88bed9770e02ed8b33343744955549d6b55a3bd6dac28ed9a3336642a40"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.573012 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tgbrt" event={"ID":"337f295d-6c74-4850-bb14-125192de4385","Type":"ContainerStarted","Data":"8381279476e92624c58b3ea09f34af46e6f5c8c4061b9534c63f0d9d299c6df3"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.573050 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tgbrt" event={"ID":"337f295d-6c74-4850-bb14-125192de4385","Type":"ContainerStarted","Data":"43e164f17fcf543c727f127b9a2bab378a33f398a4a95311b501ec89ed31ad3b"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.573449 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-tgbrt" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.573448 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-9m2sk" podStartSLOduration=116.572698478 podStartE2EDuration="1m56.572698478s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:16.494949273 +0000 UTC m=+138.516934878" watchObservedRunningTime="2025-11-23 06:47:16.572698478 +0000 UTC m=+138.594684091" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.574320 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" podStartSLOduration=117.574309627 podStartE2EDuration="1m57.574309627s" podCreationTimestamp="2025-11-23 06:45:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:16.567757856 +0000 UTC m=+138.589743480" watchObservedRunningTime="2025-11-23 06:47:16.574309627 +0000 UTC m=+138.596295241" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.582460 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lktfh" event={"ID":"50e7a859-2c04-4eff-889d-633ced116f34","Type":"ContainerStarted","Data":"2ba02e4871c19ab64a86c0f626b850dac3564b2e3f7ce65ac1bdc2d5d090213a"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.582530 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lktfh" event={"ID":"50e7a859-2c04-4eff-889d-633ced116f34","Type":"ContainerStarted","Data":"24bc6dadf6fc505dcb819f0e65717f0165fed4dbe6559eec644e2596b12c437f"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.583349 4559 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-tgbrt container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.583394 4559 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-tgbrt" podUID="337f295d-6c74-4850-bb14-125192de4385" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.584710 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vbcd9" event={"ID":"f4525676-ff36-42d8-b6de-408f2c799e37","Type":"ContainerStarted","Data":"373ff90fa9608bddbcab1925af502079a0055959a6c02c1520d26cb6c1eba726"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.610384 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ggt5f" event={"ID":"35b44360-64aa-45b1-9141-15d5f69b384a","Type":"ContainerStarted","Data":"ce42e3a638b1f8c140afd8e9678c465b12b7bf236579a2b726b376454426cfcd"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.642985 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:16 crc kubenswrapper[4559]: E1123 06:47:16.643847 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:17.143826669 +0000 UTC m=+139.165812283 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.655820 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-q5d8b" event={"ID":"7f90213c-484b-4f9d-b064-363887589530","Type":"ContainerStarted","Data":"8d3cbcb908217144b860e43608b877b1303526c778898ff4e3147857fdb92a8a"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.667391 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8xfrw" event={"ID":"a130a936-3db2-4455-bb63-fcea37eda82c","Type":"ContainerStarted","Data":"15dc75bbf07421755540f2f66618e9d52256ccfcc0018db6db400ef21dc48383"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.669165 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-78zq2" event={"ID":"833dd52a-3337-431a-af83-d75240e6ab5c","Type":"ContainerStarted","Data":"51602a03df8e2623feb717b1ecda97ec0259ccd8fa768b53689ad7e72bdfc2d2"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.669197 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-78zq2" event={"ID":"833dd52a-3337-431a-af83-d75240e6ab5c","Type":"ContainerStarted","Data":"f7e0c1e820ba384707789ec33f044eb4328ad5ba9886798e55e938c4037afc9a"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.669795 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-78zq2" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.684552 4559 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-78zq2 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.19:5443/healthz\": dial tcp 10.217.0.19:5443: connect: connection refused" start-of-body= Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.684589 4559 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-78zq2" podUID="833dd52a-3337-431a-af83-d75240e6ab5c" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.19:5443/healthz\": dial tcp 10.217.0.19:5443: connect: connection refused" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.688720 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" event={"ID":"e814799e-e374-4ad8-96a2-46e4d5defc2f","Type":"ContainerStarted","Data":"9ce45ca60ce1add8ceade7c860a805704662c9738546f7d769697a72af07bded"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.694834 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t2rfc" event={"ID":"09d28a3a-5b25-4e31-90d6-17d9319cd11c","Type":"ContainerStarted","Data":"8d6ef6457387010a9f299ab581eebec825db0c2f32270c8808d30dff1fad904e"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.694870 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t2rfc" event={"ID":"09d28a3a-5b25-4e31-90d6-17d9319cd11c","Type":"ContainerStarted","Data":"ab4872c279f316d22f44bfa5e54043eee7b74e09a268baca0534faa406395e55"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.704072 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fjv2r" podStartSLOduration=116.704061447 podStartE2EDuration="1m56.704061447s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:16.703526481 +0000 UTC m=+138.725512095" watchObservedRunningTime="2025-11-23 06:47:16.704061447 +0000 UTC m=+138.726047061" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.706268 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xwb9l" event={"ID":"1b7acf5d-cae8-429a-8661-ec3abcf056c0","Type":"ContainerStarted","Data":"83c73481fdbce66ebf5e9ce6feff42e7a72854f228ee710fd476ba1df2c3291e"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.726036 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-dlj2h" event={"ID":"16e9fcf3-adc7-4af9-a625-125ff496cfcf","Type":"ContainerStarted","Data":"b1bd8915ef29def2b815c02fb25a1694a017fbb2570b08ba9b34b89b50b6287b"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.744674 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dnz8k" podStartSLOduration=116.744658694 podStartE2EDuration="1m56.744658694s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:16.743130532 +0000 UTC m=+138.765116146" watchObservedRunningTime="2025-11-23 06:47:16.744658694 +0000 UTC m=+138.766644308" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.746387 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:16 crc kubenswrapper[4559]: E1123 06:47:16.748192 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:17.248178317 +0000 UTC m=+139.270163931 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.764657 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-ztnsp" podStartSLOduration=116.764628625 podStartE2EDuration="1m56.764628625s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:16.762932764 +0000 UTC m=+138.784918378" watchObservedRunningTime="2025-11-23 06:47:16.764628625 +0000 UTC m=+138.786614239" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.766943 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-xvd4d" event={"ID":"6796a9bd-a428-4b0a-8723-d8a6a252de84","Type":"ContainerStarted","Data":"b62a236308102438a80cc28535cfdc1e01fd091cc74122207037decdbb62735c"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.807714 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9775t" event={"ID":"266931fd-4205-4e31-9ded-9882a2a41921","Type":"ContainerStarted","Data":"2c541216d392ad53e1181dd3167d7f2e6831cda166c7f69ae9fae1a7f6f81179"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.807751 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9775t" event={"ID":"266931fd-4205-4e31-9ded-9882a2a41921","Type":"ContainerStarted","Data":"6de439dc0e333365caffa68e440eb19a8e734e7e183f46cdca2310c863254d26"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.810186 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-b2l6v" event={"ID":"be99ba37-d853-42ee-bf91-102daae86795","Type":"ContainerStarted","Data":"c3ffb5ff67e78835cf8e98b5f9bf7798bbbcc2626b4b2d10c7a675bcd1a121d1"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.810212 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-b2l6v" event={"ID":"be99ba37-d853-42ee-bf91-102daae86795","Type":"ContainerStarted","Data":"9badaa0638121ce4faae4a0c3f9f38790ed25ab512d6f101b373c6ec78eee7a5"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.823007 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vbcd9" podStartSLOduration=116.822989383 podStartE2EDuration="1m56.822989383s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:16.822013049 +0000 UTC m=+138.843998653" watchObservedRunningTime="2025-11-23 06:47:16.822989383 +0000 UTC m=+138.844974996" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.823459 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mkb95" event={"ID":"0ac1ccfb-fbd3-4de6-83bd-ff1ddcd531eb","Type":"ContainerStarted","Data":"48aa7c48d4df1d49b79ffbd2c8dddb12e6302088c7fc441464c7d00361d00c3b"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.824307 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mkb95" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.825426 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zr8cm" podStartSLOduration=116.825417594 podStartE2EDuration="1m56.825417594s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:16.784171263 +0000 UTC m=+138.806156877" watchObservedRunningTime="2025-11-23 06:47:16.825417594 +0000 UTC m=+138.847403207" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.830724 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2p9ch" event={"ID":"254348bf-91ce-448f-8bb6-c4f6f72bde3f","Type":"ContainerStarted","Data":"7df4105f445208d66bce8e67a8de20577553367eb585a914cc400bb9dc96c276"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.842387 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-8wt84" event={"ID":"a91fbd9b-8264-4ac4-b1bd-d62143b7a4d2","Type":"ContainerStarted","Data":"4ce54e1358803c540dbec6fcb67df442c9320e5d67ac78f50b0d71601a12d0ea"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.842410 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-8wt84" event={"ID":"a91fbd9b-8264-4ac4-b1bd-d62143b7a4d2","Type":"ContainerStarted","Data":"bebe75b4783701fe5e7b12f0a2881ec5946c3a8a57fa7b8b1e71b88d442b4f39"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.842459 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-78zq2" podStartSLOduration=116.842448202 podStartE2EDuration="1m56.842448202s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:16.840912236 +0000 UTC m=+138.862897850" watchObservedRunningTime="2025-11-23 06:47:16.842448202 +0000 UTC m=+138.864433816" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.851795 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:16 crc kubenswrapper[4559]: E1123 06:47:16.853312 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:17.35329604 +0000 UTC m=+139.375281654 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.868359 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xwb9l" podStartSLOduration=116.868348604 podStartE2EDuration="1m56.868348604s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:16.867820952 +0000 UTC m=+138.889806565" watchObservedRunningTime="2025-11-23 06:47:16.868348604 +0000 UTC m=+138.890334218" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.874823 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-nnldq" event={"ID":"d1d0d430-031f-448e-9763-69deb17abc4a","Type":"ContainerStarted","Data":"fb8c83c3dc960f919ee7c61b2b0f2954d8b86e86288f9ed98a5ac3db1f7fdc4e"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.875524 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mkb95" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.882593 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2j2z7" event={"ID":"ae746e9c-3187-4a3b-a439-b5ef25b37caf","Type":"ContainerStarted","Data":"01c72d1005f7aec0ed92112e7b64bd3aac1f54c29c6546c3b9c2f1f4f4842c62"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.882622 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2j2z7" event={"ID":"ae746e9c-3187-4a3b-a439-b5ef25b37caf","Type":"ContainerStarted","Data":"235fc3280207a24865fa9e681557f59d556a76614fa4f112a3e496af82cee1c2"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.886281 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t2rfc" podStartSLOduration=116.886271085 podStartE2EDuration="1m56.886271085s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:16.884837493 +0000 UTC m=+138.906823107" watchObservedRunningTime="2025-11-23 06:47:16.886271085 +0000 UTC m=+138.908256699" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.900503 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j665r" event={"ID":"58f094f8-8c16-4929-ae94-e52241c5b7d1","Type":"ContainerStarted","Data":"ea14d3080b9e99d5921439d4940d5a1b8d6ed437b4ec8a5ed7563302bc046de2"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.919542 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lktfh" podStartSLOduration=116.919526542 podStartE2EDuration="1m56.919526542s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:16.91827707 +0000 UTC m=+138.940262684" watchObservedRunningTime="2025-11-23 06:47:16.919526542 +0000 UTC m=+138.941512157" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.924205 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-ltkhr" event={"ID":"528f6e87-f9e3-46f4-b40b-127b4092b8aa","Type":"ContainerStarted","Data":"76469bfcf50aef929280e5b4685bba522b339f7661f93610b84cd33f5ab4fd3f"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.924244 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-ltkhr" event={"ID":"528f6e87-f9e3-46f4-b40b-127b4092b8aa","Type":"ContainerStarted","Data":"94418e3e3e13a15eecb10d94181ff39aec50964b14c1228d912580e8b8445d90"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.931865 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8d5rf" event={"ID":"5fc8b0b0-d23e-4493-b6f6-6b3036036f39","Type":"ContainerStarted","Data":"98d779550e1d0beb5669debb3312fec5ee3ae2be71e91f44602a023fe56b1cbb"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.931901 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8d5rf" event={"ID":"5fc8b0b0-d23e-4493-b6f6-6b3036036f39","Type":"ContainerStarted","Data":"888eca76b2e9f6cb2c6d5cd6e693882e0918e1ba7dc8010b1c68c3f14eb48935"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.932584 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8d5rf" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.932856 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-dlj2h" podStartSLOduration=116.932844782 podStartE2EDuration="1m56.932844782s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:16.932779939 +0000 UTC m=+138.954765553" watchObservedRunningTime="2025-11-23 06:47:16.932844782 +0000 UTC m=+138.954830395" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.941025 4559 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-8d5rf container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" start-of-body= Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.941064 4559 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8d5rf" podUID="5fc8b0b0-d23e-4493-b6f6-6b3036036f39" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.955863 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:16 crc kubenswrapper[4559]: E1123 06:47:16.956223 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:17.456211892 +0000 UTC m=+139.478197506 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.958737 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-tgbrt" podStartSLOduration=116.958726207 podStartE2EDuration="1m56.958726207s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:16.956678608 +0000 UTC m=+138.978664223" watchObservedRunningTime="2025-11-23 06:47:16.958726207 +0000 UTC m=+138.980711821" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.963359 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-9gwbh" event={"ID":"a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba","Type":"ContainerStarted","Data":"abcafab5c5ae9a323d0b08210e0d1d585e0024498c0451460bcf523cf829e661"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.974687 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-mfl2b" event={"ID":"624ec98f-bf3b-4651-9d9a-26e6a4377140","Type":"ContainerStarted","Data":"e5ce083d70b5bedf4038a11807e541bb03c37a580b0f0419e9fd3fa8ef88f621"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.979533 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qt4fl" event={"ID":"80d890b8-c797-4392-a48a-7894059147d4","Type":"ContainerStarted","Data":"0061492e560fb4ea282018dfb45d9e7b9e1011f083dba2e4bcaa6024fc633a44"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.979782 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qt4fl" event={"ID":"80d890b8-c797-4392-a48a-7894059147d4","Type":"ContainerStarted","Data":"f8e84e797a5cff7745974e41b07434d4fe880c30f146a9e2838f782c05cfb22a"} Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.980766 4559 patch_prober.go:28] interesting pod/downloads-7954f5f757-p9rps container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.980790 4559 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-p9rps" podUID="4316762b-b889-41c0-aa31-f874bc0fa3fe" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Nov 23 06:47:16 crc kubenswrapper[4559]: I1123 06:47:16.998692 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" podStartSLOduration=116.998679914 podStartE2EDuration="1m56.998679914s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:16.994720565 +0000 UTC m=+139.016706180" watchObservedRunningTime="2025-11-23 06:47:16.998679914 +0000 UTC m=+139.020665527" Nov 23 06:47:17 crc kubenswrapper[4559]: I1123 06:47:17.055267 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8xfrw" podStartSLOduration=117.055250493 podStartE2EDuration="1m57.055250493s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:17.024542182 +0000 UTC m=+139.046527795" watchObservedRunningTime="2025-11-23 06:47:17.055250493 +0000 UTC m=+139.077236106" Nov 23 06:47:17 crc kubenswrapper[4559]: I1123 06:47:17.060731 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:17 crc kubenswrapper[4559]: E1123 06:47:17.062401 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:17.562387134 +0000 UTC m=+139.584372748 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:17 crc kubenswrapper[4559]: I1123 06:47:17.109082 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-9gwbh" Nov 23 06:47:17 crc kubenswrapper[4559]: I1123 06:47:17.112176 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2j2z7" podStartSLOduration=117.112161478 podStartE2EDuration="1m57.112161478s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:17.1119625 +0000 UTC m=+139.133948113" watchObservedRunningTime="2025-11-23 06:47:17.112161478 +0000 UTC m=+139.134147082" Nov 23 06:47:17 crc kubenswrapper[4559]: I1123 06:47:17.112465 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-8wt84" podStartSLOduration=117.11245807 podStartE2EDuration="1m57.11245807s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:17.054836476 +0000 UTC m=+139.076822090" watchObservedRunningTime="2025-11-23 06:47:17.11245807 +0000 UTC m=+139.134443685" Nov 23 06:47:17 crc kubenswrapper[4559]: I1123 06:47:17.127816 4559 patch_prober.go:28] interesting pod/router-default-5444994796-9gwbh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:47:17 crc kubenswrapper[4559]: [-]has-synced failed: reason withheld Nov 23 06:47:17 crc kubenswrapper[4559]: [+]process-running ok Nov 23 06:47:17 crc kubenswrapper[4559]: healthz check failed Nov 23 06:47:17 crc kubenswrapper[4559]: I1123 06:47:17.127853 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9gwbh" podUID="a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:47:17 crc kubenswrapper[4559]: I1123 06:47:17.137829 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-ltkhr" podStartSLOduration=117.137815831 podStartE2EDuration="1m57.137815831s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:17.136502257 +0000 UTC m=+139.158487871" watchObservedRunningTime="2025-11-23 06:47:17.137815831 +0000 UTC m=+139.159801445" Nov 23 06:47:17 crc kubenswrapper[4559]: I1123 06:47:17.163202 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:17 crc kubenswrapper[4559]: E1123 06:47:17.163588 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:17.6635784 +0000 UTC m=+139.685564015 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:17 crc kubenswrapper[4559]: I1123 06:47:17.220196 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8d5rf" podStartSLOduration=117.220179978 podStartE2EDuration="1m57.220179978s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:17.169213631 +0000 UTC m=+139.191199245" watchObservedRunningTime="2025-11-23 06:47:17.220179978 +0000 UTC m=+139.242165693" Nov 23 06:47:17 crc kubenswrapper[4559]: I1123 06:47:17.266791 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mkb95" podStartSLOduration=117.266773824 podStartE2EDuration="1m57.266773824s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:17.244188016 +0000 UTC m=+139.266173630" watchObservedRunningTime="2025-11-23 06:47:17.266773824 +0000 UTC m=+139.288759437" Nov 23 06:47:17 crc kubenswrapper[4559]: I1123 06:47:17.275217 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:17 crc kubenswrapper[4559]: E1123 06:47:17.275568 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:17.775553885 +0000 UTC m=+139.797539490 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:17 crc kubenswrapper[4559]: I1123 06:47:17.317589 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-b2l6v" podStartSLOduration=6.317572994 podStartE2EDuration="6.317572994s" podCreationTimestamp="2025-11-23 06:47:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:17.316627578 +0000 UTC m=+139.338613193" watchObservedRunningTime="2025-11-23 06:47:17.317572994 +0000 UTC m=+139.339558608" Nov 23 06:47:17 crc kubenswrapper[4559]: I1123 06:47:17.318455 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9775t" podStartSLOduration=117.318448436 podStartE2EDuration="1m57.318448436s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:17.268274524 +0000 UTC m=+139.290260138" watchObservedRunningTime="2025-11-23 06:47:17.318448436 +0000 UTC m=+139.340434050" Nov 23 06:47:17 crc kubenswrapper[4559]: I1123 06:47:17.353741 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-9gwbh" podStartSLOduration=117.353725653 podStartE2EDuration="1m57.353725653s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:17.351561703 +0000 UTC m=+139.373547317" watchObservedRunningTime="2025-11-23 06:47:17.353725653 +0000 UTC m=+139.375711266" Nov 23 06:47:17 crc kubenswrapper[4559]: I1123 06:47:17.366916 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qt4fl" podStartSLOduration=117.366901821 podStartE2EDuration="1m57.366901821s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:17.36657401 +0000 UTC m=+139.388559623" watchObservedRunningTime="2025-11-23 06:47:17.366901821 +0000 UTC m=+139.388887435" Nov 23 06:47:17 crc kubenswrapper[4559]: I1123 06:47:17.376903 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:17 crc kubenswrapper[4559]: E1123 06:47:17.377206 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:17.877181642 +0000 UTC m=+139.899167256 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:17 crc kubenswrapper[4559]: I1123 06:47:17.416676 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-nnldq" podStartSLOduration=118.41666248 podStartE2EDuration="1m58.41666248s" podCreationTimestamp="2025-11-23 06:45:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:17.414964035 +0000 UTC m=+139.436949649" watchObservedRunningTime="2025-11-23 06:47:17.41666248 +0000 UTC m=+139.438648093" Nov 23 06:47:17 crc kubenswrapper[4559]: I1123 06:47:17.477219 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:17 crc kubenswrapper[4559]: E1123 06:47:17.477506 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:17.977493309 +0000 UTC m=+139.999478923 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:17 crc kubenswrapper[4559]: I1123 06:47:17.578028 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:17 crc kubenswrapper[4559]: E1123 06:47:17.578296 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:18.078283474 +0000 UTC m=+140.100269088 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:17 crc kubenswrapper[4559]: I1123 06:47:17.678501 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:17 crc kubenswrapper[4559]: E1123 06:47:17.678765 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:18.178751608 +0000 UTC m=+140.200737223 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:17 crc kubenswrapper[4559]: I1123 06:47:17.779241 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:17 crc kubenswrapper[4559]: E1123 06:47:17.779543 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:18.279532025 +0000 UTC m=+140.301517640 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:17 crc kubenswrapper[4559]: I1123 06:47:17.879998 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:17 crc kubenswrapper[4559]: E1123 06:47:17.880234 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:18.38021135 +0000 UTC m=+140.402196964 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:17 crc kubenswrapper[4559]: I1123 06:47:17.880488 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:17 crc kubenswrapper[4559]: E1123 06:47:17.880730 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:18.380719455 +0000 UTC m=+140.402705069 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:17 crc kubenswrapper[4559]: I1123 06:47:17.981201 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:17 crc kubenswrapper[4559]: E1123 06:47:17.981336 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:18.481317927 +0000 UTC m=+140.503303541 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:17 crc kubenswrapper[4559]: I1123 06:47:17.981468 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:17 crc kubenswrapper[4559]: E1123 06:47:17.981737 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:18.481726984 +0000 UTC m=+140.503712598 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.014098 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2p9ch" event={"ID":"254348bf-91ce-448f-8bb6-c4f6f72bde3f","Type":"ContainerStarted","Data":"0b792470a164a50bf44f556b26189efdc99864d5733b9bfd5e229c37392990ff"} Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.014133 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2p9ch" event={"ID":"254348bf-91ce-448f-8bb6-c4f6f72bde3f","Type":"ContainerStarted","Data":"d8e31eee74259658ac04cfce4d3d19eb0a236eb6b0ddeef50eb0470a5f909b19"} Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.015755 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r779q" event={"ID":"04abe9e5-6ffe-429d-9274-73d133fd0d52","Type":"ContainerStarted","Data":"b43a2db6cf254d7412d98c5c1ecb24994cdb476c5c611ddfd0e501af04a84d48"} Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.015819 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r779q" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.025759 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" event={"ID":"994432aa-8794-4818-94bd-ce28d97c7163","Type":"ContainerStarted","Data":"b4dd8ebce39793dd093f2a230403865def6767810554d07894c0030b8e38b958"} Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.025787 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" event={"ID":"994432aa-8794-4818-94bd-ce28d97c7163","Type":"ContainerStarted","Data":"e8f9c8b084411b2662e8ea5c093a3b2d71306ca51f9ca1c04f3669de038f5bad"} Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.028087 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t2rfc" event={"ID":"09d28a3a-5b25-4e31-90d6-17d9319cd11c","Type":"ContainerStarted","Data":"2cc0a37007d98ede5d8de066f6a4c4244103d40e1dd8a97b0670f44f34753acd"} Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.029596 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vbcd9" event={"ID":"f4525676-ff36-42d8-b6de-408f2c799e37","Type":"ContainerStarted","Data":"37bc3bb3086cdd79e3eb1b3f1a178c6e1bd86b800b9817cf78a2aeab5e4d0e26"} Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.033362 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-mfl2b" event={"ID":"624ec98f-bf3b-4651-9d9a-26e6a4377140","Type":"ContainerStarted","Data":"fd502abc43659d18d22f925b259b76f9d4772a1fbaea01aab908ff02e297749a"} Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.033386 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-mfl2b" event={"ID":"624ec98f-bf3b-4651-9d9a-26e6a4377140","Type":"ContainerStarted","Data":"e4a1b748d3cc40715af2c15f32641306c7c81950d6458d66b847857e3823e4ae"} Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.033398 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-mfl2b" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.042845 4559 patch_prober.go:28] interesting pod/downloads-7954f5f757-p9rps container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.042890 4559 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-p9rps" podUID="4316762b-b889-41c0-aa31-f874bc0fa3fe" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.044264 4559 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-tgbrt container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.044305 4559 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-tgbrt" podUID="337f295d-6c74-4850-bb14-125192de4385" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.044551 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r779q" podStartSLOduration=118.044537581 podStartE2EDuration="1m58.044537581s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:18.043519869 +0000 UTC m=+140.065505483" watchObservedRunningTime="2025-11-23 06:47:18.044537581 +0000 UTC m=+140.066523195" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.057633 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8d5rf" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.068726 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-mfl2b" podStartSLOduration=7.068717866 podStartE2EDuration="7.068717866s" podCreationTimestamp="2025-11-23 06:47:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:18.067408881 +0000 UTC m=+140.089394496" watchObservedRunningTime="2025-11-23 06:47:18.068717866 +0000 UTC m=+140.090703481" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.078633 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qhvl6"] Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.079363 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qhvl6" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.080833 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.085049 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:18 crc kubenswrapper[4559]: E1123 06:47:18.085291 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:18.585279905 +0000 UTC m=+140.607265510 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.103778 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qhvl6"] Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.106237 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" podStartSLOduration=119.106223334 podStartE2EDuration="1m59.106223334s" podCreationTimestamp="2025-11-23 06:45:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:18.105495883 +0000 UTC m=+140.127481497" watchObservedRunningTime="2025-11-23 06:47:18.106223334 +0000 UTC m=+140.128208947" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.111764 4559 patch_prober.go:28] interesting pod/router-default-5444994796-9gwbh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:47:18 crc kubenswrapper[4559]: [-]has-synced failed: reason withheld Nov 23 06:47:18 crc kubenswrapper[4559]: [+]process-running ok Nov 23 06:47:18 crc kubenswrapper[4559]: healthz check failed Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.111799 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9gwbh" podUID="a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.187328 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.187626 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/680dc567-9fb8-4396-adb4-76bb44250c9c-utilities\") pod \"certified-operators-qhvl6\" (UID: \"680dc567-9fb8-4396-adb4-76bb44250c9c\") " pod="openshift-marketplace/certified-operators-qhvl6" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.187684 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dktv\" (UniqueName: \"kubernetes.io/projected/680dc567-9fb8-4396-adb4-76bb44250c9c-kube-api-access-7dktv\") pod \"certified-operators-qhvl6\" (UID: \"680dc567-9fb8-4396-adb4-76bb44250c9c\") " pod="openshift-marketplace/certified-operators-qhvl6" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.187997 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/680dc567-9fb8-4396-adb4-76bb44250c9c-catalog-content\") pod \"certified-operators-qhvl6\" (UID: \"680dc567-9fb8-4396-adb4-76bb44250c9c\") " pod="openshift-marketplace/certified-operators-qhvl6" Nov 23 06:47:18 crc kubenswrapper[4559]: E1123 06:47:18.190430 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:18.690415923 +0000 UTC m=+140.712401536 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.289319 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.289571 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/680dc567-9fb8-4396-adb4-76bb44250c9c-utilities\") pod \"certified-operators-qhvl6\" (UID: \"680dc567-9fb8-4396-adb4-76bb44250c9c\") " pod="openshift-marketplace/certified-operators-qhvl6" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.289595 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dktv\" (UniqueName: \"kubernetes.io/projected/680dc567-9fb8-4396-adb4-76bb44250c9c-kube-api-access-7dktv\") pod \"certified-operators-qhvl6\" (UID: \"680dc567-9fb8-4396-adb4-76bb44250c9c\") " pod="openshift-marketplace/certified-operators-qhvl6" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.289634 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/680dc567-9fb8-4396-adb4-76bb44250c9c-catalog-content\") pod \"certified-operators-qhvl6\" (UID: \"680dc567-9fb8-4396-adb4-76bb44250c9c\") " pod="openshift-marketplace/certified-operators-qhvl6" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.290015 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/680dc567-9fb8-4396-adb4-76bb44250c9c-catalog-content\") pod \"certified-operators-qhvl6\" (UID: \"680dc567-9fb8-4396-adb4-76bb44250c9c\") " pod="openshift-marketplace/certified-operators-qhvl6" Nov 23 06:47:18 crc kubenswrapper[4559]: E1123 06:47:18.290079 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:18.790067195 +0000 UTC m=+140.812052810 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.290275 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/680dc567-9fb8-4396-adb4-76bb44250c9c-utilities\") pod \"certified-operators-qhvl6\" (UID: \"680dc567-9fb8-4396-adb4-76bb44250c9c\") " pod="openshift-marketplace/certified-operators-qhvl6" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.306524 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-whrv6"] Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.307303 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-whrv6"] Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.307378 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-whrv6" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.324571 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.325031 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.325957 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dktv\" (UniqueName: \"kubernetes.io/projected/680dc567-9fb8-4396-adb4-76bb44250c9c-kube-api-access-7dktv\") pod \"certified-operators-qhvl6\" (UID: \"680dc567-9fb8-4396-adb4-76bb44250c9c\") " pod="openshift-marketplace/certified-operators-qhvl6" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.361705 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.361934 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.391949 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:18 crc kubenswrapper[4559]: E1123 06:47:18.392229 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:18.892216683 +0000 UTC m=+140.914202297 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.392379 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qhvl6" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.392734 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.410682 4559 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.475071 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nbmqd"] Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.475908 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nbmqd" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.492405 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.492579 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d1c86c0-8a6f-4914-9593-20ca1220ed8b-catalog-content\") pod \"community-operators-whrv6\" (UID: \"9d1c86c0-8a6f-4914-9593-20ca1220ed8b\") " pod="openshift-marketplace/community-operators-whrv6" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.492621 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d1c86c0-8a6f-4914-9593-20ca1220ed8b-utilities\") pod \"community-operators-whrv6\" (UID: \"9d1c86c0-8a6f-4914-9593-20ca1220ed8b\") " pod="openshift-marketplace/community-operators-whrv6" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.492738 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gh72\" (UniqueName: \"kubernetes.io/projected/9d1c86c0-8a6f-4914-9593-20ca1220ed8b-kube-api-access-9gh72\") pod \"community-operators-whrv6\" (UID: \"9d1c86c0-8a6f-4914-9593-20ca1220ed8b\") " pod="openshift-marketplace/community-operators-whrv6" Nov 23 06:47:18 crc kubenswrapper[4559]: E1123 06:47:18.493068 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:18.993052745 +0000 UTC m=+141.015038359 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.498771 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nbmqd"] Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.503487 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-78zq2" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.602683 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7848f717-c7e3-40df-a4d8-854794e3c2e8-utilities\") pod \"certified-operators-nbmqd\" (UID: \"7848f717-c7e3-40df-a4d8-854794e3c2e8\") " pod="openshift-marketplace/certified-operators-nbmqd" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.602907 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.602936 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gh72\" (UniqueName: \"kubernetes.io/projected/9d1c86c0-8a6f-4914-9593-20ca1220ed8b-kube-api-access-9gh72\") pod \"community-operators-whrv6\" (UID: \"9d1c86c0-8a6f-4914-9593-20ca1220ed8b\") " pod="openshift-marketplace/community-operators-whrv6" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.602972 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7848f717-c7e3-40df-a4d8-854794e3c2e8-catalog-content\") pod \"certified-operators-nbmqd\" (UID: \"7848f717-c7e3-40df-a4d8-854794e3c2e8\") " pod="openshift-marketplace/certified-operators-nbmqd" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.603028 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqxbv\" (UniqueName: \"kubernetes.io/projected/7848f717-c7e3-40df-a4d8-854794e3c2e8-kube-api-access-dqxbv\") pod \"certified-operators-nbmqd\" (UID: \"7848f717-c7e3-40df-a4d8-854794e3c2e8\") " pod="openshift-marketplace/certified-operators-nbmqd" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.603103 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d1c86c0-8a6f-4914-9593-20ca1220ed8b-catalog-content\") pod \"community-operators-whrv6\" (UID: \"9d1c86c0-8a6f-4914-9593-20ca1220ed8b\") " pod="openshift-marketplace/community-operators-whrv6" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.603131 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d1c86c0-8a6f-4914-9593-20ca1220ed8b-utilities\") pod \"community-operators-whrv6\" (UID: \"9d1c86c0-8a6f-4914-9593-20ca1220ed8b\") " pod="openshift-marketplace/community-operators-whrv6" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.603476 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d1c86c0-8a6f-4914-9593-20ca1220ed8b-utilities\") pod \"community-operators-whrv6\" (UID: \"9d1c86c0-8a6f-4914-9593-20ca1220ed8b\") " pod="openshift-marketplace/community-operators-whrv6" Nov 23 06:47:18 crc kubenswrapper[4559]: E1123 06:47:18.603725 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:19.103714446 +0000 UTC m=+141.125700060 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.604480 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d1c86c0-8a6f-4914-9593-20ca1220ed8b-catalog-content\") pod \"community-operators-whrv6\" (UID: \"9d1c86c0-8a6f-4914-9593-20ca1220ed8b\") " pod="openshift-marketplace/community-operators-whrv6" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.637811 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.637849 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.653485 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gh72\" (UniqueName: \"kubernetes.io/projected/9d1c86c0-8a6f-4914-9593-20ca1220ed8b-kube-api-access-9gh72\") pod \"community-operators-whrv6\" (UID: \"9d1c86c0-8a6f-4914-9593-20ca1220ed8b\") " pod="openshift-marketplace/community-operators-whrv6" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.662789 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-whrv6" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.684574 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vx7hg"] Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.686713 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vx7hg" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.704265 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.704483 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7848f717-c7e3-40df-a4d8-854794e3c2e8-utilities\") pod \"certified-operators-nbmqd\" (UID: \"7848f717-c7e3-40df-a4d8-854794e3c2e8\") " pod="openshift-marketplace/certified-operators-nbmqd" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.704529 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7848f717-c7e3-40df-a4d8-854794e3c2e8-catalog-content\") pod \"certified-operators-nbmqd\" (UID: \"7848f717-c7e3-40df-a4d8-854794e3c2e8\") " pod="openshift-marketplace/certified-operators-nbmqd" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.704576 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqxbv\" (UniqueName: \"kubernetes.io/projected/7848f717-c7e3-40df-a4d8-854794e3c2e8-kube-api-access-dqxbv\") pod \"certified-operators-nbmqd\" (UID: \"7848f717-c7e3-40df-a4d8-854794e3c2e8\") " pod="openshift-marketplace/certified-operators-nbmqd" Nov 23 06:47:18 crc kubenswrapper[4559]: E1123 06:47:18.704877 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:19.204864835 +0000 UTC m=+141.226850449 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.705223 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7848f717-c7e3-40df-a4d8-854794e3c2e8-utilities\") pod \"certified-operators-nbmqd\" (UID: \"7848f717-c7e3-40df-a4d8-854794e3c2e8\") " pod="openshift-marketplace/certified-operators-nbmqd" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.705420 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7848f717-c7e3-40df-a4d8-854794e3c2e8-catalog-content\") pod \"certified-operators-nbmqd\" (UID: \"7848f717-c7e3-40df-a4d8-854794e3c2e8\") " pod="openshift-marketplace/certified-operators-nbmqd" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.719439 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vx7hg"] Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.734442 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqxbv\" (UniqueName: \"kubernetes.io/projected/7848f717-c7e3-40df-a4d8-854794e3c2e8-kube-api-access-dqxbv\") pod \"certified-operators-nbmqd\" (UID: \"7848f717-c7e3-40df-a4d8-854794e3c2e8\") " pod="openshift-marketplace/certified-operators-nbmqd" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.779325 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qhvl6"] Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.790924 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nbmqd" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.806324 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88092487-aea4-46ae-8918-35e1adf96708-utilities\") pod \"community-operators-vx7hg\" (UID: \"88092487-aea4-46ae-8918-35e1adf96708\") " pod="openshift-marketplace/community-operators-vx7hg" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.806377 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88092487-aea4-46ae-8918-35e1adf96708-catalog-content\") pod \"community-operators-vx7hg\" (UID: \"88092487-aea4-46ae-8918-35e1adf96708\") " pod="openshift-marketplace/community-operators-vx7hg" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.806430 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq8sm\" (UniqueName: \"kubernetes.io/projected/88092487-aea4-46ae-8918-35e1adf96708-kube-api-access-kq8sm\") pod \"community-operators-vx7hg\" (UID: \"88092487-aea4-46ae-8918-35e1adf96708\") " pod="openshift-marketplace/community-operators-vx7hg" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.806464 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:18 crc kubenswrapper[4559]: E1123 06:47:18.806715 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:19.306704253 +0000 UTC m=+141.328689868 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:18 crc kubenswrapper[4559]: W1123 06:47:18.831826 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod680dc567_9fb8_4396_adb4_76bb44250c9c.slice/crio-8a25804bc3e1fcd40c2a4a9d074af5b808167846082d5116deca070c9c793622 WatchSource:0}: Error finding container 8a25804bc3e1fcd40c2a4a9d074af5b808167846082d5116deca070c9c793622: Status 404 returned error can't find the container with id 8a25804bc3e1fcd40c2a4a9d074af5b808167846082d5116deca070c9c793622 Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.907011 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:18 crc kubenswrapper[4559]: E1123 06:47:18.907522 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:19.407497444 +0000 UTC m=+141.429483059 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.907657 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88092487-aea4-46ae-8918-35e1adf96708-utilities\") pod \"community-operators-vx7hg\" (UID: \"88092487-aea4-46ae-8918-35e1adf96708\") " pod="openshift-marketplace/community-operators-vx7hg" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.907700 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88092487-aea4-46ae-8918-35e1adf96708-catalog-content\") pod \"community-operators-vx7hg\" (UID: \"88092487-aea4-46ae-8918-35e1adf96708\") " pod="openshift-marketplace/community-operators-vx7hg" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.907780 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq8sm\" (UniqueName: \"kubernetes.io/projected/88092487-aea4-46ae-8918-35e1adf96708-kube-api-access-kq8sm\") pod \"community-operators-vx7hg\" (UID: \"88092487-aea4-46ae-8918-35e1adf96708\") " pod="openshift-marketplace/community-operators-vx7hg" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.909218 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88092487-aea4-46ae-8918-35e1adf96708-utilities\") pod \"community-operators-vx7hg\" (UID: \"88092487-aea4-46ae-8918-35e1adf96708\") " pod="openshift-marketplace/community-operators-vx7hg" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.909556 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88092487-aea4-46ae-8918-35e1adf96708-catalog-content\") pod \"community-operators-vx7hg\" (UID: \"88092487-aea4-46ae-8918-35e1adf96708\") " pod="openshift-marketplace/community-operators-vx7hg" Nov 23 06:47:18 crc kubenswrapper[4559]: I1123 06:47:18.937986 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq8sm\" (UniqueName: \"kubernetes.io/projected/88092487-aea4-46ae-8918-35e1adf96708-kube-api-access-kq8sm\") pod \"community-operators-vx7hg\" (UID: \"88092487-aea4-46ae-8918-35e1adf96708\") " pod="openshift-marketplace/community-operators-vx7hg" Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.009025 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:19 crc kubenswrapper[4559]: E1123 06:47:19.009463 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:19.509449757 +0000 UTC m=+141.531435371 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.016291 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-whrv6"] Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.039707 4559 generic.go:334] "Generic (PLEG): container finished" podID="680dc567-9fb8-4396-adb4-76bb44250c9c" containerID="270ed989791553c1822cfb134bdec27290409b476d30e7e2c5e95ed1da292bb9" exitCode=0 Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.039761 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qhvl6" event={"ID":"680dc567-9fb8-4396-adb4-76bb44250c9c","Type":"ContainerDied","Data":"270ed989791553c1822cfb134bdec27290409b476d30e7e2c5e95ed1da292bb9"} Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.039784 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qhvl6" event={"ID":"680dc567-9fb8-4396-adb4-76bb44250c9c","Type":"ContainerStarted","Data":"8a25804bc3e1fcd40c2a4a9d074af5b808167846082d5116deca070c9c793622"} Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.040077 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vx7hg" Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.047414 4559 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.048890 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2p9ch" event={"ID":"254348bf-91ce-448f-8bb6-c4f6f72bde3f","Type":"ContainerStarted","Data":"31b24b0f2992386a6feb644aba8e065184c686798805dc0bca749c3ce515c8ac"} Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.048922 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2p9ch" event={"ID":"254348bf-91ce-448f-8bb6-c4f6f72bde3f","Type":"ContainerStarted","Data":"a56dceca9b74d5f78318b4538f60353121b1652d43c7cddce02ecf239a452a3c"} Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.063281 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntvtt" Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.067104 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-tgbrt" Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.107771 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-2p9ch" podStartSLOduration=8.107757619 podStartE2EDuration="8.107757619s" podCreationTimestamp="2025-11-23 06:47:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:19.088379563 +0000 UTC m=+141.110365177" watchObservedRunningTime="2025-11-23 06:47:19.107757619 +0000 UTC m=+141.129743232" Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.109042 4559 patch_prober.go:28] interesting pod/router-default-5444994796-9gwbh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:47:19 crc kubenswrapper[4559]: [-]has-synced failed: reason withheld Nov 23 06:47:19 crc kubenswrapper[4559]: [+]process-running ok Nov 23 06:47:19 crc kubenswrapper[4559]: healthz check failed Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.109083 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9gwbh" podUID="a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.110264 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:19 crc kubenswrapper[4559]: E1123 06:47:19.110573 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:19.61055507 +0000 UTC m=+141.632540684 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.111068 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:19 crc kubenswrapper[4559]: E1123 06:47:19.112604 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:19.612593812 +0000 UTC m=+141.634579426 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.207237 4559 patch_prober.go:28] interesting pod/apiserver-76f77b778f-5kfl6 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Nov 23 06:47:19 crc kubenswrapper[4559]: [+]log ok Nov 23 06:47:19 crc kubenswrapper[4559]: [+]etcd ok Nov 23 06:47:19 crc kubenswrapper[4559]: [+]poststarthook/start-apiserver-admission-initializer ok Nov 23 06:47:19 crc kubenswrapper[4559]: [+]poststarthook/generic-apiserver-start-informers ok Nov 23 06:47:19 crc kubenswrapper[4559]: [+]poststarthook/max-in-flight-filter ok Nov 23 06:47:19 crc kubenswrapper[4559]: [+]poststarthook/storage-object-count-tracker-hook ok Nov 23 06:47:19 crc kubenswrapper[4559]: [+]poststarthook/image.openshift.io-apiserver-caches ok Nov 23 06:47:19 crc kubenswrapper[4559]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Nov 23 06:47:19 crc kubenswrapper[4559]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Nov 23 06:47:19 crc kubenswrapper[4559]: [+]poststarthook/project.openshift.io-projectcache ok Nov 23 06:47:19 crc kubenswrapper[4559]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Nov 23 06:47:19 crc kubenswrapper[4559]: [-]poststarthook/openshift.io-startinformers failed: reason withheld Nov 23 06:47:19 crc kubenswrapper[4559]: [+]poststarthook/openshift.io-restmapperupdater ok Nov 23 06:47:19 crc kubenswrapper[4559]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Nov 23 06:47:19 crc kubenswrapper[4559]: livez check failed Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.207452 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" podUID="994432aa-8794-4818-94bd-ce28d97c7163" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.211897 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:19 crc kubenswrapper[4559]: E1123 06:47:19.213715 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 06:47:19.713698414 +0000 UTC m=+141.735684029 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.273863 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vx7hg"] Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.274636 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nbmqd"] Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.313462 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:19 crc kubenswrapper[4559]: E1123 06:47:19.313766 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 06:47:19.813751551 +0000 UTC m=+141.835737165 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccrq2" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.392232 4559 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-23T06:47:18.410696724Z","Handler":null,"Name":""} Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.396298 4559 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.396328 4559 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.414931 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.419136 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.516076 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.517961 4559 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.517986 4559 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.532591 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccrq2\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.549331 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.689515 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ccrq2"] Nov 23 06:47:19 crc kubenswrapper[4559]: W1123 06:47:19.693107 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2aee8d30_e0ce_4b6a_8e3c_d5e42d0bf87d.slice/crio-839794abcb333c8ab2d3709f630533953507fcd49b8f9a3eb3c80da034b03694 WatchSource:0}: Error finding container 839794abcb333c8ab2d3709f630533953507fcd49b8f9a3eb3c80da034b03694: Status 404 returned error can't find the container with id 839794abcb333c8ab2d3709f630533953507fcd49b8f9a3eb3c80da034b03694 Nov 23 06:47:19 crc kubenswrapper[4559]: I1123 06:47:19.753745 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fjv2r" Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.056061 4559 generic.go:334] "Generic (PLEG): container finished" podID="9d1c86c0-8a6f-4914-9593-20ca1220ed8b" containerID="7f65ec22cc44f19a9f306c55c9921d313d8ba7c45edfe2d63cb4f3b1dd40d5cd" exitCode=0 Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.056197 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whrv6" event={"ID":"9d1c86c0-8a6f-4914-9593-20ca1220ed8b","Type":"ContainerDied","Data":"7f65ec22cc44f19a9f306c55c9921d313d8ba7c45edfe2d63cb4f3b1dd40d5cd"} Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.056403 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whrv6" event={"ID":"9d1c86c0-8a6f-4914-9593-20ca1220ed8b","Type":"ContainerStarted","Data":"e01d714e3a17603b995e082d90b605afc5bd1651356adda9586662a3517766c7"} Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.057884 4559 generic.go:334] "Generic (PLEG): container finished" podID="d1d0d430-031f-448e-9763-69deb17abc4a" containerID="fb8c83c3dc960f919ee7c61b2b0f2954d8b86e86288f9ed98a5ac3db1f7fdc4e" exitCode=0 Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.057942 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-nnldq" event={"ID":"d1d0d430-031f-448e-9763-69deb17abc4a","Type":"ContainerDied","Data":"fb8c83c3dc960f919ee7c61b2b0f2954d8b86e86288f9ed98a5ac3db1f7fdc4e"} Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.059663 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" event={"ID":"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d","Type":"ContainerStarted","Data":"4756d4be1b1c5e4807c762f8492154b9a2f27691c5d2c91d3f70165db42be5d3"} Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.059690 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" event={"ID":"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d","Type":"ContainerStarted","Data":"839794abcb333c8ab2d3709f630533953507fcd49b8f9a3eb3c80da034b03694"} Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.059797 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.068114 4559 generic.go:334] "Generic (PLEG): container finished" podID="7848f717-c7e3-40df-a4d8-854794e3c2e8" containerID="bc6b991c3573a01c8577e700208bc288b17ebe5a7dae751940e70f16219a0aa7" exitCode=0 Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.068155 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nbmqd" event={"ID":"7848f717-c7e3-40df-a4d8-854794e3c2e8","Type":"ContainerDied","Data":"bc6b991c3573a01c8577e700208bc288b17ebe5a7dae751940e70f16219a0aa7"} Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.068202 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nbmqd" event={"ID":"7848f717-c7e3-40df-a4d8-854794e3c2e8","Type":"ContainerStarted","Data":"6959e76b09d0f5832cf28d8f4e6c10307973653e13210052a5488c75d73fae85"} Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.071490 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-m9ltm"] Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.072436 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m9ltm" Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.074523 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.075090 4559 generic.go:334] "Generic (PLEG): container finished" podID="88092487-aea4-46ae-8918-35e1adf96708" containerID="09e85832676ae7db5010b3d73350a4c2b5589ad6bcabeaf3a7db0494ca356285" exitCode=0 Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.075325 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vx7hg" event={"ID":"88092487-aea4-46ae-8918-35e1adf96708","Type":"ContainerDied","Data":"09e85832676ae7db5010b3d73350a4c2b5589ad6bcabeaf3a7db0494ca356285"} Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.075356 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vx7hg" event={"ID":"88092487-aea4-46ae-8918-35e1adf96708","Type":"ContainerStarted","Data":"7f2614485ae243f6229a0ccd0452f92235464eec30867b99d75d6cf9195d97a9"} Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.086078 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m9ltm"] Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.093692 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" podStartSLOduration=120.093679297 podStartE2EDuration="2m0.093679297s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:20.092594577 +0000 UTC m=+142.114580191" watchObservedRunningTime="2025-11-23 06:47:20.093679297 +0000 UTC m=+142.115664911" Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.107960 4559 patch_prober.go:28] interesting pod/router-default-5444994796-9gwbh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:47:20 crc kubenswrapper[4559]: [-]has-synced failed: reason withheld Nov 23 06:47:20 crc kubenswrapper[4559]: [+]process-running ok Nov 23 06:47:20 crc kubenswrapper[4559]: healthz check failed Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.107999 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9gwbh" podUID="a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.227446 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxpff\" (UniqueName: \"kubernetes.io/projected/98640d4a-0cbc-428c-8f7a-25a1b36f16f9-kube-api-access-lxpff\") pod \"redhat-marketplace-m9ltm\" (UID: \"98640d4a-0cbc-428c-8f7a-25a1b36f16f9\") " pod="openshift-marketplace/redhat-marketplace-m9ltm" Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.227522 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98640d4a-0cbc-428c-8f7a-25a1b36f16f9-catalog-content\") pod \"redhat-marketplace-m9ltm\" (UID: \"98640d4a-0cbc-428c-8f7a-25a1b36f16f9\") " pod="openshift-marketplace/redhat-marketplace-m9ltm" Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.227610 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98640d4a-0cbc-428c-8f7a-25a1b36f16f9-utilities\") pod \"redhat-marketplace-m9ltm\" (UID: \"98640d4a-0cbc-428c-8f7a-25a1b36f16f9\") " pod="openshift-marketplace/redhat-marketplace-m9ltm" Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.282632 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.331292 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxpff\" (UniqueName: \"kubernetes.io/projected/98640d4a-0cbc-428c-8f7a-25a1b36f16f9-kube-api-access-lxpff\") pod \"redhat-marketplace-m9ltm\" (UID: \"98640d4a-0cbc-428c-8f7a-25a1b36f16f9\") " pod="openshift-marketplace/redhat-marketplace-m9ltm" Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.331342 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98640d4a-0cbc-428c-8f7a-25a1b36f16f9-catalog-content\") pod \"redhat-marketplace-m9ltm\" (UID: \"98640d4a-0cbc-428c-8f7a-25a1b36f16f9\") " pod="openshift-marketplace/redhat-marketplace-m9ltm" Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.331384 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98640d4a-0cbc-428c-8f7a-25a1b36f16f9-utilities\") pod \"redhat-marketplace-m9ltm\" (UID: \"98640d4a-0cbc-428c-8f7a-25a1b36f16f9\") " pod="openshift-marketplace/redhat-marketplace-m9ltm" Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.331759 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98640d4a-0cbc-428c-8f7a-25a1b36f16f9-catalog-content\") pod \"redhat-marketplace-m9ltm\" (UID: \"98640d4a-0cbc-428c-8f7a-25a1b36f16f9\") " pod="openshift-marketplace/redhat-marketplace-m9ltm" Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.331792 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98640d4a-0cbc-428c-8f7a-25a1b36f16f9-utilities\") pod \"redhat-marketplace-m9ltm\" (UID: \"98640d4a-0cbc-428c-8f7a-25a1b36f16f9\") " pod="openshift-marketplace/redhat-marketplace-m9ltm" Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.349608 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxpff\" (UniqueName: \"kubernetes.io/projected/98640d4a-0cbc-428c-8f7a-25a1b36f16f9-kube-api-access-lxpff\") pod \"redhat-marketplace-m9ltm\" (UID: \"98640d4a-0cbc-428c-8f7a-25a1b36f16f9\") " pod="openshift-marketplace/redhat-marketplace-m9ltm" Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.388890 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m9ltm" Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.476545 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9xk28"] Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.478884 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9xk28" Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.480622 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9xk28"] Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.635017 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79f79383-e237-4748-bdc0-1edb62550872-catalog-content\") pod \"redhat-marketplace-9xk28\" (UID: \"79f79383-e237-4748-bdc0-1edb62550872\") " pod="openshift-marketplace/redhat-marketplace-9xk28" Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.635073 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9slv6\" (UniqueName: \"kubernetes.io/projected/79f79383-e237-4748-bdc0-1edb62550872-kube-api-access-9slv6\") pod \"redhat-marketplace-9xk28\" (UID: \"79f79383-e237-4748-bdc0-1edb62550872\") " pod="openshift-marketplace/redhat-marketplace-9xk28" Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.635135 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79f79383-e237-4748-bdc0-1edb62550872-utilities\") pod \"redhat-marketplace-9xk28\" (UID: \"79f79383-e237-4748-bdc0-1edb62550872\") " pod="openshift-marketplace/redhat-marketplace-9xk28" Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.735006 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m9ltm"] Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.736636 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79f79383-e237-4748-bdc0-1edb62550872-catalog-content\") pod \"redhat-marketplace-9xk28\" (UID: \"79f79383-e237-4748-bdc0-1edb62550872\") " pod="openshift-marketplace/redhat-marketplace-9xk28" Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.736713 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9slv6\" (UniqueName: \"kubernetes.io/projected/79f79383-e237-4748-bdc0-1edb62550872-kube-api-access-9slv6\") pod \"redhat-marketplace-9xk28\" (UID: \"79f79383-e237-4748-bdc0-1edb62550872\") " pod="openshift-marketplace/redhat-marketplace-9xk28" Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.736757 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79f79383-e237-4748-bdc0-1edb62550872-utilities\") pod \"redhat-marketplace-9xk28\" (UID: \"79f79383-e237-4748-bdc0-1edb62550872\") " pod="openshift-marketplace/redhat-marketplace-9xk28" Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.737302 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79f79383-e237-4748-bdc0-1edb62550872-catalog-content\") pod \"redhat-marketplace-9xk28\" (UID: \"79f79383-e237-4748-bdc0-1edb62550872\") " pod="openshift-marketplace/redhat-marketplace-9xk28" Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.737388 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79f79383-e237-4748-bdc0-1edb62550872-utilities\") pod \"redhat-marketplace-9xk28\" (UID: \"79f79383-e237-4748-bdc0-1edb62550872\") " pod="openshift-marketplace/redhat-marketplace-9xk28" Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.751418 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9slv6\" (UniqueName: \"kubernetes.io/projected/79f79383-e237-4748-bdc0-1edb62550872-kube-api-access-9slv6\") pod \"redhat-marketplace-9xk28\" (UID: \"79f79383-e237-4748-bdc0-1edb62550872\") " pod="openshift-marketplace/redhat-marketplace-9xk28" Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.791991 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9xk28" Nov 23 06:47:20 crc kubenswrapper[4559]: I1123 06:47:20.942822 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9xk28"] Nov 23 06:47:20 crc kubenswrapper[4559]: W1123 06:47:20.955053 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod79f79383_e237_4748_bdc0_1edb62550872.slice/crio-3f705a6d15152768eb7af2a5502224b0f6e9ac81fcc9e0ec454f6b8a7dec31a4 WatchSource:0}: Error finding container 3f705a6d15152768eb7af2a5502224b0f6e9ac81fcc9e0ec454f6b8a7dec31a4: Status 404 returned error can't find the container with id 3f705a6d15152768eb7af2a5502224b0f6e9ac81fcc9e0ec454f6b8a7dec31a4 Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.080426 4559 generic.go:334] "Generic (PLEG): container finished" podID="98640d4a-0cbc-428c-8f7a-25a1b36f16f9" containerID="a7938a6b9af471f28df9be4c1aad4fedc267f6c68e3d8df79902a4c8d7100fe0" exitCode=0 Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.080480 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m9ltm" event={"ID":"98640d4a-0cbc-428c-8f7a-25a1b36f16f9","Type":"ContainerDied","Data":"a7938a6b9af471f28df9be4c1aad4fedc267f6c68e3d8df79902a4c8d7100fe0"} Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.080504 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m9ltm" event={"ID":"98640d4a-0cbc-428c-8f7a-25a1b36f16f9","Type":"ContainerStarted","Data":"c48ff1f993a90e1623d868ab760bb058ed436429c40f6a9910351988e590c75b"} Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.085780 4559 generic.go:334] "Generic (PLEG): container finished" podID="79f79383-e237-4748-bdc0-1edb62550872" containerID="1b4be2cec437fdc1bc4a6614ef3fe251bb7a7c8e410dd4f477165e2778c74f08" exitCode=0 Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.085851 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xk28" event={"ID":"79f79383-e237-4748-bdc0-1edb62550872","Type":"ContainerDied","Data":"1b4be2cec437fdc1bc4a6614ef3fe251bb7a7c8e410dd4f477165e2778c74f08"} Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.087418 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xk28" event={"ID":"79f79383-e237-4748-bdc0-1edb62550872","Type":"ContainerStarted","Data":"3f705a6d15152768eb7af2a5502224b0f6e9ac81fcc9e0ec454f6b8a7dec31a4"} Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.104776 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.105302 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.109385 4559 patch_prober.go:28] interesting pod/router-default-5444994796-9gwbh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:47:21 crc kubenswrapper[4559]: [-]has-synced failed: reason withheld Nov 23 06:47:21 crc kubenswrapper[4559]: [+]process-running ok Nov 23 06:47:21 crc kubenswrapper[4559]: healthz check failed Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.109419 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9gwbh" podUID="a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.111149 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.111349 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.116224 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.242543 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/21b0a5f3-85fe-4f27-a139-3f1d21b063fa-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"21b0a5f3-85fe-4f27-a139-3f1d21b063fa\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.242719 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/21b0a5f3-85fe-4f27-a139-3f1d21b063fa-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"21b0a5f3-85fe-4f27-a139-3f1d21b063fa\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.273246 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rz5lj"] Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.274427 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rz5lj" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.277548 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.279276 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rz5lj"] Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.343957 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/21b0a5f3-85fe-4f27-a139-3f1d21b063fa-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"21b0a5f3-85fe-4f27-a139-3f1d21b063fa\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.344245 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/21b0a5f3-85fe-4f27-a139-3f1d21b063fa-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"21b0a5f3-85fe-4f27-a139-3f1d21b063fa\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.344114 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/21b0a5f3-85fe-4f27-a139-3f1d21b063fa-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"21b0a5f3-85fe-4f27-a139-3f1d21b063fa\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.359948 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/21b0a5f3-85fe-4f27-a139-3f1d21b063fa-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"21b0a5f3-85fe-4f27-a139-3f1d21b063fa\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.424428 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.445497 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmnvz\" (UniqueName: \"kubernetes.io/projected/f59cbebb-100b-4d79-83f8-bf1a625caf93-kube-api-access-vmnvz\") pod \"redhat-operators-rz5lj\" (UID: \"f59cbebb-100b-4d79-83f8-bf1a625caf93\") " pod="openshift-marketplace/redhat-operators-rz5lj" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.445592 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f59cbebb-100b-4d79-83f8-bf1a625caf93-utilities\") pod \"redhat-operators-rz5lj\" (UID: \"f59cbebb-100b-4d79-83f8-bf1a625caf93\") " pod="openshift-marketplace/redhat-operators-rz5lj" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.445613 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f59cbebb-100b-4d79-83f8-bf1a625caf93-catalog-content\") pod \"redhat-operators-rz5lj\" (UID: \"f59cbebb-100b-4d79-83f8-bf1a625caf93\") " pod="openshift-marketplace/redhat-operators-rz5lj" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.547346 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmnvz\" (UniqueName: \"kubernetes.io/projected/f59cbebb-100b-4d79-83f8-bf1a625caf93-kube-api-access-vmnvz\") pod \"redhat-operators-rz5lj\" (UID: \"f59cbebb-100b-4d79-83f8-bf1a625caf93\") " pod="openshift-marketplace/redhat-operators-rz5lj" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.547447 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f59cbebb-100b-4d79-83f8-bf1a625caf93-utilities\") pod \"redhat-operators-rz5lj\" (UID: \"f59cbebb-100b-4d79-83f8-bf1a625caf93\") " pod="openshift-marketplace/redhat-operators-rz5lj" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.547466 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f59cbebb-100b-4d79-83f8-bf1a625caf93-catalog-content\") pod \"redhat-operators-rz5lj\" (UID: \"f59cbebb-100b-4d79-83f8-bf1a625caf93\") " pod="openshift-marketplace/redhat-operators-rz5lj" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.547962 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f59cbebb-100b-4d79-83f8-bf1a625caf93-utilities\") pod \"redhat-operators-rz5lj\" (UID: \"f59cbebb-100b-4d79-83f8-bf1a625caf93\") " pod="openshift-marketplace/redhat-operators-rz5lj" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.547992 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f59cbebb-100b-4d79-83f8-bf1a625caf93-catalog-content\") pod \"redhat-operators-rz5lj\" (UID: \"f59cbebb-100b-4d79-83f8-bf1a625caf93\") " pod="openshift-marketplace/redhat-operators-rz5lj" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.573117 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmnvz\" (UniqueName: \"kubernetes.io/projected/f59cbebb-100b-4d79-83f8-bf1a625caf93-kube-api-access-vmnvz\") pod \"redhat-operators-rz5lj\" (UID: \"f59cbebb-100b-4d79-83f8-bf1a625caf93\") " pod="openshift-marketplace/redhat-operators-rz5lj" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.589696 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rz5lj" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.672242 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vqk9p"] Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.673112 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vqk9p" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.679294 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vqk9p"] Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.852722 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25720f84-faae-4d25-8bac-7942852c0a28-catalog-content\") pod \"redhat-operators-vqk9p\" (UID: \"25720f84-faae-4d25-8bac-7942852c0a28\") " pod="openshift-marketplace/redhat-operators-vqk9p" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.852963 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5psk\" (UniqueName: \"kubernetes.io/projected/25720f84-faae-4d25-8bac-7942852c0a28-kube-api-access-c5psk\") pod \"redhat-operators-vqk9p\" (UID: \"25720f84-faae-4d25-8bac-7942852c0a28\") " pod="openshift-marketplace/redhat-operators-vqk9p" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.853131 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25720f84-faae-4d25-8bac-7942852c0a28-utilities\") pod \"redhat-operators-vqk9p\" (UID: \"25720f84-faae-4d25-8bac-7942852c0a28\") " pod="openshift-marketplace/redhat-operators-vqk9p" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.954559 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5psk\" (UniqueName: \"kubernetes.io/projected/25720f84-faae-4d25-8bac-7942852c0a28-kube-api-access-c5psk\") pod \"redhat-operators-vqk9p\" (UID: \"25720f84-faae-4d25-8bac-7942852c0a28\") " pod="openshift-marketplace/redhat-operators-vqk9p" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.954627 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25720f84-faae-4d25-8bac-7942852c0a28-utilities\") pod \"redhat-operators-vqk9p\" (UID: \"25720f84-faae-4d25-8bac-7942852c0a28\") " pod="openshift-marketplace/redhat-operators-vqk9p" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.954681 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25720f84-faae-4d25-8bac-7942852c0a28-catalog-content\") pod \"redhat-operators-vqk9p\" (UID: \"25720f84-faae-4d25-8bac-7942852c0a28\") " pod="openshift-marketplace/redhat-operators-vqk9p" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.955275 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25720f84-faae-4d25-8bac-7942852c0a28-catalog-content\") pod \"redhat-operators-vqk9p\" (UID: \"25720f84-faae-4d25-8bac-7942852c0a28\") " pod="openshift-marketplace/redhat-operators-vqk9p" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.955530 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25720f84-faae-4d25-8bac-7942852c0a28-utilities\") pod \"redhat-operators-vqk9p\" (UID: \"25720f84-faae-4d25-8bac-7942852c0a28\") " pod="openshift-marketplace/redhat-operators-vqk9p" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.971953 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5psk\" (UniqueName: \"kubernetes.io/projected/25720f84-faae-4d25-8bac-7942852c0a28-kube-api-access-c5psk\") pod \"redhat-operators-vqk9p\" (UID: \"25720f84-faae-4d25-8bac-7942852c0a28\") " pod="openshift-marketplace/redhat-operators-vqk9p" Nov 23 06:47:21 crc kubenswrapper[4559]: I1123 06:47:21.989367 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vqk9p" Nov 23 06:47:22 crc kubenswrapper[4559]: I1123 06:47:22.106828 4559 patch_prober.go:28] interesting pod/router-default-5444994796-9gwbh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:47:22 crc kubenswrapper[4559]: [-]has-synced failed: reason withheld Nov 23 06:47:22 crc kubenswrapper[4559]: [+]process-running ok Nov 23 06:47:22 crc kubenswrapper[4559]: healthz check failed Nov 23 06:47:22 crc kubenswrapper[4559]: I1123 06:47:22.106864 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9gwbh" podUID="a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:47:23 crc kubenswrapper[4559]: I1123 06:47:23.124700 4559 patch_prober.go:28] interesting pod/router-default-5444994796-9gwbh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:47:23 crc kubenswrapper[4559]: [-]has-synced failed: reason withheld Nov 23 06:47:23 crc kubenswrapper[4559]: [+]process-running ok Nov 23 06:47:23 crc kubenswrapper[4559]: healthz check failed Nov 23 06:47:23 crc kubenswrapper[4559]: I1123 06:47:23.124956 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9gwbh" podUID="a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:47:23 crc kubenswrapper[4559]: I1123 06:47:23.463271 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-p9rps" Nov 23 06:47:23 crc kubenswrapper[4559]: I1123 06:47:23.518026 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 23 06:47:23 crc kubenswrapper[4559]: I1123 06:47:23.518941 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 06:47:23 crc kubenswrapper[4559]: I1123 06:47:23.520924 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 23 06:47:23 crc kubenswrapper[4559]: I1123 06:47:23.521116 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 23 06:47:23 crc kubenswrapper[4559]: I1123 06:47:23.525007 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 23 06:47:23 crc kubenswrapper[4559]: I1123 06:47:23.637966 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:23 crc kubenswrapper[4559]: I1123 06:47:23.641612 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-5kfl6" Nov 23 06:47:23 crc kubenswrapper[4559]: I1123 06:47:23.691359 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eccd095a-81da-4c51-8107-0ad84257c648-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"eccd095a-81da-4c51-8107-0ad84257c648\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 06:47:23 crc kubenswrapper[4559]: I1123 06:47:23.692201 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/eccd095a-81da-4c51-8107-0ad84257c648-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"eccd095a-81da-4c51-8107-0ad84257c648\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 06:47:23 crc kubenswrapper[4559]: I1123 06:47:23.736361 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:47:23 crc kubenswrapper[4559]: I1123 06:47:23.736415 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:47:23 crc kubenswrapper[4559]: I1123 06:47:23.738370 4559 patch_prober.go:28] interesting pod/console-f9d7485db-xvd4d container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Nov 23 06:47:23 crc kubenswrapper[4559]: I1123 06:47:23.738409 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-xvd4d" podUID="6796a9bd-a428-4b0a-8723-d8a6a252de84" containerName="console" probeResult="failure" output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" Nov 23 06:47:23 crc kubenswrapper[4559]: I1123 06:47:23.793546 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eccd095a-81da-4c51-8107-0ad84257c648-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"eccd095a-81da-4c51-8107-0ad84257c648\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 06:47:23 crc kubenswrapper[4559]: I1123 06:47:23.793818 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/eccd095a-81da-4c51-8107-0ad84257c648-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"eccd095a-81da-4c51-8107-0ad84257c648\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 06:47:23 crc kubenswrapper[4559]: I1123 06:47:23.793922 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/eccd095a-81da-4c51-8107-0ad84257c648-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"eccd095a-81da-4c51-8107-0ad84257c648\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 06:47:23 crc kubenswrapper[4559]: I1123 06:47:23.808458 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eccd095a-81da-4c51-8107-0ad84257c648-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"eccd095a-81da-4c51-8107-0ad84257c648\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 06:47:23 crc kubenswrapper[4559]: I1123 06:47:23.840874 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 06:47:23 crc kubenswrapper[4559]: I1123 06:47:23.968732 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-nnldq" Nov 23 06:47:24 crc kubenswrapper[4559]: I1123 06:47:24.000338 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzshk\" (UniqueName: \"kubernetes.io/projected/d1d0d430-031f-448e-9763-69deb17abc4a-kube-api-access-nzshk\") pod \"d1d0d430-031f-448e-9763-69deb17abc4a\" (UID: \"d1d0d430-031f-448e-9763-69deb17abc4a\") " Nov 23 06:47:24 crc kubenswrapper[4559]: I1123 06:47:24.000376 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d1d0d430-031f-448e-9763-69deb17abc4a-secret-volume\") pod \"d1d0d430-031f-448e-9763-69deb17abc4a\" (UID: \"d1d0d430-031f-448e-9763-69deb17abc4a\") " Nov 23 06:47:24 crc kubenswrapper[4559]: I1123 06:47:24.000496 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d1d0d430-031f-448e-9763-69deb17abc4a-config-volume\") pod \"d1d0d430-031f-448e-9763-69deb17abc4a\" (UID: \"d1d0d430-031f-448e-9763-69deb17abc4a\") " Nov 23 06:47:24 crc kubenswrapper[4559]: I1123 06:47:24.000767 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:47:24 crc kubenswrapper[4559]: I1123 06:47:24.003056 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1d0d430-031f-448e-9763-69deb17abc4a-config-volume" (OuterVolumeSpecName: "config-volume") pod "d1d0d430-031f-448e-9763-69deb17abc4a" (UID: "d1d0d430-031f-448e-9763-69deb17abc4a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:47:24 crc kubenswrapper[4559]: I1123 06:47:24.004966 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1d0d430-031f-448e-9763-69deb17abc4a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d1d0d430-031f-448e-9763-69deb17abc4a" (UID: "d1d0d430-031f-448e-9763-69deb17abc4a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:47:24 crc kubenswrapper[4559]: I1123 06:47:24.007001 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1d0d430-031f-448e-9763-69deb17abc4a-kube-api-access-nzshk" (OuterVolumeSpecName: "kube-api-access-nzshk") pod "d1d0d430-031f-448e-9763-69deb17abc4a" (UID: "d1d0d430-031f-448e-9763-69deb17abc4a"). InnerVolumeSpecName "kube-api-access-nzshk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:47:24 crc kubenswrapper[4559]: I1123 06:47:24.010360 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:47:24 crc kubenswrapper[4559]: I1123 06:47:24.102694 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:47:24 crc kubenswrapper[4559]: I1123 06:47:24.102806 4559 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d1d0d430-031f-448e-9763-69deb17abc4a-config-volume\") on node \"crc\" DevicePath \"\"" Nov 23 06:47:24 crc kubenswrapper[4559]: I1123 06:47:24.102831 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzshk\" (UniqueName: \"kubernetes.io/projected/d1d0d430-031f-448e-9763-69deb17abc4a-kube-api-access-nzshk\") on node \"crc\" DevicePath \"\"" Nov 23 06:47:24 crc kubenswrapper[4559]: I1123 06:47:24.102843 4559 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d1d0d430-031f-448e-9763-69deb17abc4a-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 23 06:47:24 crc kubenswrapper[4559]: I1123 06:47:24.105345 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-9gwbh" Nov 23 06:47:24 crc kubenswrapper[4559]: I1123 06:47:24.106845 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:47:24 crc kubenswrapper[4559]: I1123 06:47:24.108472 4559 patch_prober.go:28] interesting pod/router-default-5444994796-9gwbh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:47:24 crc kubenswrapper[4559]: [-]has-synced failed: reason withheld Nov 23 06:47:24 crc kubenswrapper[4559]: [+]process-running ok Nov 23 06:47:24 crc kubenswrapper[4559]: healthz check failed Nov 23 06:47:24 crc kubenswrapper[4559]: I1123 06:47:24.108508 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9gwbh" podUID="a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:47:24 crc kubenswrapper[4559]: I1123 06:47:24.119464 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-nnldq" Nov 23 06:47:24 crc kubenswrapper[4559]: I1123 06:47:24.120050 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398005-nnldq" event={"ID":"d1d0d430-031f-448e-9763-69deb17abc4a","Type":"ContainerDied","Data":"e947b16875ec3b22595f0769f7b02bd01da5259a48b5d893ca321c2a3df3db1c"} Nov 23 06:47:24 crc kubenswrapper[4559]: I1123 06:47:24.120092 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e947b16875ec3b22595f0769f7b02bd01da5259a48b5d893ca321c2a3df3db1c" Nov 23 06:47:24 crc kubenswrapper[4559]: I1123 06:47:24.203717 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:47:24 crc kubenswrapper[4559]: I1123 06:47:24.203793 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:47:24 crc kubenswrapper[4559]: I1123 06:47:24.205729 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:47:24 crc kubenswrapper[4559]: I1123 06:47:24.215724 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:47:24 crc kubenswrapper[4559]: I1123 06:47:24.282603 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:47:24 crc kubenswrapper[4559]: I1123 06:47:24.287494 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 06:47:24 crc kubenswrapper[4559]: I1123 06:47:24.384442 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 06:47:24 crc kubenswrapper[4559]: I1123 06:47:24.974148 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rz5lj"] Nov 23 06:47:24 crc kubenswrapper[4559]: I1123 06:47:24.976250 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 23 06:47:25 crc kubenswrapper[4559]: I1123 06:47:25.069758 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vqk9p"] Nov 23 06:47:25 crc kubenswrapper[4559]: I1123 06:47:25.072331 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 23 06:47:25 crc kubenswrapper[4559]: W1123 06:47:25.075426 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podeccd095a_81da_4c51_8107_0ad84257c648.slice/crio-f1862f0210d7f73193aab3825305e5c3f759b931e5ffc7bbe8e2c909cd8f8442 WatchSource:0}: Error finding container f1862f0210d7f73193aab3825305e5c3f759b931e5ffc7bbe8e2c909cd8f8442: Status 404 returned error can't find the container with id f1862f0210d7f73193aab3825305e5c3f759b931e5ffc7bbe8e2c909cd8f8442 Nov 23 06:47:25 crc kubenswrapper[4559]: W1123 06:47:25.085896 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25720f84_faae_4d25_8bac_7942852c0a28.slice/crio-3a78d83385f885c7edaab01d9bfa79634ee9de09d9874ba478617d73e8cbe073 WatchSource:0}: Error finding container 3a78d83385f885c7edaab01d9bfa79634ee9de09d9874ba478617d73e8cbe073: Status 404 returned error can't find the container with id 3a78d83385f885c7edaab01d9bfa79634ee9de09d9874ba478617d73e8cbe073 Nov 23 06:47:25 crc kubenswrapper[4559]: W1123 06:47:25.091748 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-2afd33ea7437c3fdd57d15ab7cbde4e97b7d40a57ac23619ad641c7e22766223 WatchSource:0}: Error finding container 2afd33ea7437c3fdd57d15ab7cbde4e97b7d40a57ac23619ad641c7e22766223: Status 404 returned error can't find the container with id 2afd33ea7437c3fdd57d15ab7cbde4e97b7d40a57ac23619ad641c7e22766223 Nov 23 06:47:25 crc kubenswrapper[4559]: I1123 06:47:25.119418 4559 patch_prober.go:28] interesting pod/router-default-5444994796-9gwbh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:47:25 crc kubenswrapper[4559]: [-]has-synced failed: reason withheld Nov 23 06:47:25 crc kubenswrapper[4559]: [+]process-running ok Nov 23 06:47:25 crc kubenswrapper[4559]: healthz check failed Nov 23 06:47:25 crc kubenswrapper[4559]: I1123 06:47:25.119475 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9gwbh" podUID="a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:47:25 crc kubenswrapper[4559]: I1123 06:47:25.135272 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vqk9p" event={"ID":"25720f84-faae-4d25-8bac-7942852c0a28","Type":"ContainerStarted","Data":"3a78d83385f885c7edaab01d9bfa79634ee9de09d9874ba478617d73e8cbe073"} Nov 23 06:47:25 crc kubenswrapper[4559]: I1123 06:47:25.137203 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"2afd33ea7437c3fdd57d15ab7cbde4e97b7d40a57ac23619ad641c7e22766223"} Nov 23 06:47:25 crc kubenswrapper[4559]: I1123 06:47:25.138424 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"21b0a5f3-85fe-4f27-a139-3f1d21b063fa","Type":"ContainerStarted","Data":"8ac72cf08321c2551396100b1dca9b28574922358ec9850c08af4c03e5ce1fc9"} Nov 23 06:47:25 crc kubenswrapper[4559]: I1123 06:47:25.142304 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rz5lj" event={"ID":"f59cbebb-100b-4d79-83f8-bf1a625caf93","Type":"ContainerStarted","Data":"8f11e6da0278ec4a8b7586a1d8d6df08e9f796ba7b0314b6043d34a838b88db9"} Nov 23 06:47:25 crc kubenswrapper[4559]: I1123 06:47:25.145803 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"7398212d16f47eea1b4893e40da9fcceb4f23153992092c64f442ea39ab6c8ea"} Nov 23 06:47:25 crc kubenswrapper[4559]: I1123 06:47:25.146961 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"eccd095a-81da-4c51-8107-0ad84257c648","Type":"ContainerStarted","Data":"f1862f0210d7f73193aab3825305e5c3f759b931e5ffc7bbe8e2c909cd8f8442"} Nov 23 06:47:25 crc kubenswrapper[4559]: I1123 06:47:25.149023 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"e57bda2c47a436953b65dec48468b18af94d8781f1c3688eef8a5744f385056b"} Nov 23 06:47:26 crc kubenswrapper[4559]: I1123 06:47:26.107555 4559 patch_prober.go:28] interesting pod/router-default-5444994796-9gwbh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:47:26 crc kubenswrapper[4559]: [-]has-synced failed: reason withheld Nov 23 06:47:26 crc kubenswrapper[4559]: [+]process-running ok Nov 23 06:47:26 crc kubenswrapper[4559]: healthz check failed Nov 23 06:47:26 crc kubenswrapper[4559]: I1123 06:47:26.108068 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9gwbh" podUID="a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:47:26 crc kubenswrapper[4559]: I1123 06:47:26.166548 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 06:47:26 crc kubenswrapper[4559]: I1123 06:47:26.166591 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 06:47:26 crc kubenswrapper[4559]: I1123 06:47:26.175335 4559 generic.go:334] "Generic (PLEG): container finished" podID="eccd095a-81da-4c51-8107-0ad84257c648" containerID="911b013202bdae22f8b3ea7c02f5aaf594c78cbef99163502979dc907f1be61d" exitCode=0 Nov 23 06:47:26 crc kubenswrapper[4559]: I1123 06:47:26.175401 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"eccd095a-81da-4c51-8107-0ad84257c648","Type":"ContainerDied","Data":"911b013202bdae22f8b3ea7c02f5aaf594c78cbef99163502979dc907f1be61d"} Nov 23 06:47:26 crc kubenswrapper[4559]: I1123 06:47:26.177737 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"553eece675c2b31c1647f5849348ce3df443eaf428dd501d41141ea8b428d778"} Nov 23 06:47:26 crc kubenswrapper[4559]: I1123 06:47:26.177896 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:47:26 crc kubenswrapper[4559]: I1123 06:47:26.179339 4559 generic.go:334] "Generic (PLEG): container finished" podID="25720f84-faae-4d25-8bac-7942852c0a28" containerID="b57b779dbe3d6ca04aac0f593a45a01c1098b6c056c2e1fe1f91cd5e5bb53073" exitCode=0 Nov 23 06:47:26 crc kubenswrapper[4559]: I1123 06:47:26.179357 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vqk9p" event={"ID":"25720f84-faae-4d25-8bac-7942852c0a28","Type":"ContainerDied","Data":"b57b779dbe3d6ca04aac0f593a45a01c1098b6c056c2e1fe1f91cd5e5bb53073"} Nov 23 06:47:26 crc kubenswrapper[4559]: I1123 06:47:26.181713 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"c9a419b0800bdec748f3efa808576e5f625560042377409c862e1fef59edfa69"} Nov 23 06:47:26 crc kubenswrapper[4559]: I1123 06:47:26.183590 4559 generic.go:334] "Generic (PLEG): container finished" podID="21b0a5f3-85fe-4f27-a139-3f1d21b063fa" containerID="20fb7f7e6084717d5101a7030294a9dfd97e77083c9a6573c94e02aa4d3dc09f" exitCode=0 Nov 23 06:47:26 crc kubenswrapper[4559]: I1123 06:47:26.183684 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"21b0a5f3-85fe-4f27-a139-3f1d21b063fa","Type":"ContainerDied","Data":"20fb7f7e6084717d5101a7030294a9dfd97e77083c9a6573c94e02aa4d3dc09f"} Nov 23 06:47:26 crc kubenswrapper[4559]: I1123 06:47:26.186378 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"673d6857e443403c11d6ee9c4c56fa6fffd132515e2c2956f3fa3509e38d5ae0"} Nov 23 06:47:26 crc kubenswrapper[4559]: I1123 06:47:26.191672 4559 generic.go:334] "Generic (PLEG): container finished" podID="f59cbebb-100b-4d79-83f8-bf1a625caf93" containerID="751894a83450d6c28894578e2c2c26826c5da6a0d6bce77bb13bc2ed45643811" exitCode=0 Nov 23 06:47:26 crc kubenswrapper[4559]: I1123 06:47:26.191709 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rz5lj" event={"ID":"f59cbebb-100b-4d79-83f8-bf1a625caf93","Type":"ContainerDied","Data":"751894a83450d6c28894578e2c2c26826c5da6a0d6bce77bb13bc2ed45643811"} Nov 23 06:47:26 crc kubenswrapper[4559]: I1123 06:47:26.587305 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-mfl2b" Nov 23 06:47:27 crc kubenswrapper[4559]: I1123 06:47:27.107495 4559 patch_prober.go:28] interesting pod/router-default-5444994796-9gwbh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:47:27 crc kubenswrapper[4559]: [-]has-synced failed: reason withheld Nov 23 06:47:27 crc kubenswrapper[4559]: [+]process-running ok Nov 23 06:47:27 crc kubenswrapper[4559]: healthz check failed Nov 23 06:47:27 crc kubenswrapper[4559]: I1123 06:47:27.107747 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9gwbh" podUID="a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:47:27 crc kubenswrapper[4559]: I1123 06:47:27.455440 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 06:47:27 crc kubenswrapper[4559]: I1123 06:47:27.492085 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 06:47:27 crc kubenswrapper[4559]: I1123 06:47:27.567240 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/21b0a5f3-85fe-4f27-a139-3f1d21b063fa-kube-api-access\") pod \"21b0a5f3-85fe-4f27-a139-3f1d21b063fa\" (UID: \"21b0a5f3-85fe-4f27-a139-3f1d21b063fa\") " Nov 23 06:47:27 crc kubenswrapper[4559]: I1123 06:47:27.567296 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eccd095a-81da-4c51-8107-0ad84257c648-kube-api-access\") pod \"eccd095a-81da-4c51-8107-0ad84257c648\" (UID: \"eccd095a-81da-4c51-8107-0ad84257c648\") " Nov 23 06:47:27 crc kubenswrapper[4559]: I1123 06:47:27.567384 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/21b0a5f3-85fe-4f27-a139-3f1d21b063fa-kubelet-dir\") pod \"21b0a5f3-85fe-4f27-a139-3f1d21b063fa\" (UID: \"21b0a5f3-85fe-4f27-a139-3f1d21b063fa\") " Nov 23 06:47:27 crc kubenswrapper[4559]: I1123 06:47:27.567470 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/eccd095a-81da-4c51-8107-0ad84257c648-kubelet-dir\") pod \"eccd095a-81da-4c51-8107-0ad84257c648\" (UID: \"eccd095a-81da-4c51-8107-0ad84257c648\") " Nov 23 06:47:27 crc kubenswrapper[4559]: I1123 06:47:27.567514 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eccd095a-81da-4c51-8107-0ad84257c648-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "eccd095a-81da-4c51-8107-0ad84257c648" (UID: "eccd095a-81da-4c51-8107-0ad84257c648"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 06:47:27 crc kubenswrapper[4559]: I1123 06:47:27.567514 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/21b0a5f3-85fe-4f27-a139-3f1d21b063fa-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "21b0a5f3-85fe-4f27-a139-3f1d21b063fa" (UID: "21b0a5f3-85fe-4f27-a139-3f1d21b063fa"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 06:47:27 crc kubenswrapper[4559]: I1123 06:47:27.567785 4559 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/eccd095a-81da-4c51-8107-0ad84257c648-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 23 06:47:27 crc kubenswrapper[4559]: I1123 06:47:27.567801 4559 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/21b0a5f3-85fe-4f27-a139-3f1d21b063fa-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 23 06:47:27 crc kubenswrapper[4559]: I1123 06:47:27.573192 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eccd095a-81da-4c51-8107-0ad84257c648-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "eccd095a-81da-4c51-8107-0ad84257c648" (UID: "eccd095a-81da-4c51-8107-0ad84257c648"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:47:27 crc kubenswrapper[4559]: I1123 06:47:27.583817 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21b0a5f3-85fe-4f27-a139-3f1d21b063fa-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "21b0a5f3-85fe-4f27-a139-3f1d21b063fa" (UID: "21b0a5f3-85fe-4f27-a139-3f1d21b063fa"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:47:27 crc kubenswrapper[4559]: I1123 06:47:27.669084 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/21b0a5f3-85fe-4f27-a139-3f1d21b063fa-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 23 06:47:27 crc kubenswrapper[4559]: I1123 06:47:27.669108 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eccd095a-81da-4c51-8107-0ad84257c648-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 23 06:47:28 crc kubenswrapper[4559]: I1123 06:47:28.107049 4559 patch_prober.go:28] interesting pod/router-default-5444994796-9gwbh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:47:28 crc kubenswrapper[4559]: [-]has-synced failed: reason withheld Nov 23 06:47:28 crc kubenswrapper[4559]: [+]process-running ok Nov 23 06:47:28 crc kubenswrapper[4559]: healthz check failed Nov 23 06:47:28 crc kubenswrapper[4559]: I1123 06:47:28.107297 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9gwbh" podUID="a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:47:28 crc kubenswrapper[4559]: I1123 06:47:28.208216 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"21b0a5f3-85fe-4f27-a139-3f1d21b063fa","Type":"ContainerDied","Data":"8ac72cf08321c2551396100b1dca9b28574922358ec9850c08af4c03e5ce1fc9"} Nov 23 06:47:28 crc kubenswrapper[4559]: I1123 06:47:28.208252 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ac72cf08321c2551396100b1dca9b28574922358ec9850c08af4c03e5ce1fc9" Nov 23 06:47:28 crc kubenswrapper[4559]: I1123 06:47:28.208227 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 06:47:28 crc kubenswrapper[4559]: I1123 06:47:28.209960 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"eccd095a-81da-4c51-8107-0ad84257c648","Type":"ContainerDied","Data":"f1862f0210d7f73193aab3825305e5c3f759b931e5ffc7bbe8e2c909cd8f8442"} Nov 23 06:47:28 crc kubenswrapper[4559]: I1123 06:47:28.209999 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 06:47:28 crc kubenswrapper[4559]: I1123 06:47:28.210003 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1862f0210d7f73193aab3825305e5c3f759b931e5ffc7bbe8e2c909cd8f8442" Nov 23 06:47:29 crc kubenswrapper[4559]: I1123 06:47:29.106763 4559 patch_prober.go:28] interesting pod/router-default-5444994796-9gwbh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:47:29 crc kubenswrapper[4559]: [-]has-synced failed: reason withheld Nov 23 06:47:29 crc kubenswrapper[4559]: [+]process-running ok Nov 23 06:47:29 crc kubenswrapper[4559]: healthz check failed Nov 23 06:47:29 crc kubenswrapper[4559]: I1123 06:47:29.106805 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9gwbh" podUID="a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:47:30 crc kubenswrapper[4559]: I1123 06:47:30.108318 4559 patch_prober.go:28] interesting pod/router-default-5444994796-9gwbh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:47:30 crc kubenswrapper[4559]: [-]has-synced failed: reason withheld Nov 23 06:47:30 crc kubenswrapper[4559]: [+]process-running ok Nov 23 06:47:30 crc kubenswrapper[4559]: healthz check failed Nov 23 06:47:30 crc kubenswrapper[4559]: I1123 06:47:30.108375 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9gwbh" podUID="a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:47:31 crc kubenswrapper[4559]: I1123 06:47:31.106834 4559 patch_prober.go:28] interesting pod/router-default-5444994796-9gwbh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 06:47:31 crc kubenswrapper[4559]: [-]has-synced failed: reason withheld Nov 23 06:47:31 crc kubenswrapper[4559]: [+]process-running ok Nov 23 06:47:31 crc kubenswrapper[4559]: healthz check failed Nov 23 06:47:31 crc kubenswrapper[4559]: I1123 06:47:31.107408 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9gwbh" podUID="a21ee2f0-e1ca-4d2a-aa2e-069dc5eae1ba" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 06:47:32 crc kubenswrapper[4559]: I1123 06:47:32.106933 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-9gwbh" Nov 23 06:47:32 crc kubenswrapper[4559]: I1123 06:47:32.108821 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-9gwbh" Nov 23 06:47:33 crc kubenswrapper[4559]: I1123 06:47:33.736131 4559 patch_prober.go:28] interesting pod/console-f9d7485db-xvd4d container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Nov 23 06:47:33 crc kubenswrapper[4559]: I1123 06:47:33.736389 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-xvd4d" podUID="6796a9bd-a428-4b0a-8723-d8a6a252de84" containerName="console" probeResult="failure" output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" Nov 23 06:47:38 crc kubenswrapper[4559]: I1123 06:47:38.247602 4559 generic.go:334] "Generic (PLEG): container finished" podID="88092487-aea4-46ae-8918-35e1adf96708" containerID="196d20033d13965be63bc2490bdb1a7109970eb0f96f0ebc2195ce983ae494da" exitCode=0 Nov 23 06:47:38 crc kubenswrapper[4559]: I1123 06:47:38.247672 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vx7hg" event={"ID":"88092487-aea4-46ae-8918-35e1adf96708","Type":"ContainerDied","Data":"196d20033d13965be63bc2490bdb1a7109970eb0f96f0ebc2195ce983ae494da"} Nov 23 06:47:38 crc kubenswrapper[4559]: I1123 06:47:38.249229 4559 generic.go:334] "Generic (PLEG): container finished" podID="9d1c86c0-8a6f-4914-9593-20ca1220ed8b" containerID="3fbae69bb1331f980e3710ca233be9b9d5bf936ee95a62c9a0598863aed0654f" exitCode=0 Nov 23 06:47:38 crc kubenswrapper[4559]: I1123 06:47:38.249255 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whrv6" event={"ID":"9d1c86c0-8a6f-4914-9593-20ca1220ed8b","Type":"ContainerDied","Data":"3fbae69bb1331f980e3710ca233be9b9d5bf936ee95a62c9a0598863aed0654f"} Nov 23 06:47:38 crc kubenswrapper[4559]: I1123 06:47:38.250890 4559 generic.go:334] "Generic (PLEG): container finished" podID="25720f84-faae-4d25-8bac-7942852c0a28" containerID="c9b71830cd1aa9010ffd3021cff2c654949bf62d94c1757afad72aeae26d025f" exitCode=0 Nov 23 06:47:38 crc kubenswrapper[4559]: I1123 06:47:38.250947 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vqk9p" event={"ID":"25720f84-faae-4d25-8bac-7942852c0a28","Type":"ContainerDied","Data":"c9b71830cd1aa9010ffd3021cff2c654949bf62d94c1757afad72aeae26d025f"} Nov 23 06:47:38 crc kubenswrapper[4559]: I1123 06:47:38.252732 4559 generic.go:334] "Generic (PLEG): container finished" podID="680dc567-9fb8-4396-adb4-76bb44250c9c" containerID="b06e34d44c1451202da39a5592c3d490ef6b02a8f069a17d1fc1f4329c38022a" exitCode=0 Nov 23 06:47:38 crc kubenswrapper[4559]: I1123 06:47:38.252784 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qhvl6" event={"ID":"680dc567-9fb8-4396-adb4-76bb44250c9c","Type":"ContainerDied","Data":"b06e34d44c1451202da39a5592c3d490ef6b02a8f069a17d1fc1f4329c38022a"} Nov 23 06:47:38 crc kubenswrapper[4559]: I1123 06:47:38.253902 4559 generic.go:334] "Generic (PLEG): container finished" podID="f59cbebb-100b-4d79-83f8-bf1a625caf93" containerID="cfd840e030d50c634bf5c50ec46181610c44cf82ed4ac54e427cc91d197fc0e6" exitCode=0 Nov 23 06:47:38 crc kubenswrapper[4559]: I1123 06:47:38.253933 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rz5lj" event={"ID":"f59cbebb-100b-4d79-83f8-bf1a625caf93","Type":"ContainerDied","Data":"cfd840e030d50c634bf5c50ec46181610c44cf82ed4ac54e427cc91d197fc0e6"} Nov 23 06:47:38 crc kubenswrapper[4559]: I1123 06:47:38.257128 4559 generic.go:334] "Generic (PLEG): container finished" podID="7848f717-c7e3-40df-a4d8-854794e3c2e8" containerID="ab1ed9a1bf45ad6d3d3347dff288a670fcf80fbcf3b24cd1b0931a98faeec8d6" exitCode=0 Nov 23 06:47:38 crc kubenswrapper[4559]: I1123 06:47:38.257181 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nbmqd" event={"ID":"7848f717-c7e3-40df-a4d8-854794e3c2e8","Type":"ContainerDied","Data":"ab1ed9a1bf45ad6d3d3347dff288a670fcf80fbcf3b24cd1b0931a98faeec8d6"} Nov 23 06:47:38 crc kubenswrapper[4559]: I1123 06:47:38.258885 4559 generic.go:334] "Generic (PLEG): container finished" podID="98640d4a-0cbc-428c-8f7a-25a1b36f16f9" containerID="ab8cc8db23b649332b5cfa20be82533ab922e3841569a311ea5d8f5111ea1296" exitCode=0 Nov 23 06:47:38 crc kubenswrapper[4559]: I1123 06:47:38.258934 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m9ltm" event={"ID":"98640d4a-0cbc-428c-8f7a-25a1b36f16f9","Type":"ContainerDied","Data":"ab8cc8db23b649332b5cfa20be82533ab922e3841569a311ea5d8f5111ea1296"} Nov 23 06:47:38 crc kubenswrapper[4559]: I1123 06:47:38.262345 4559 generic.go:334] "Generic (PLEG): container finished" podID="79f79383-e237-4748-bdc0-1edb62550872" containerID="6d86471e497eda6109bf5d54d2f2ddb90797a88381ebb9c485fe92531868670f" exitCode=0 Nov 23 06:47:38 crc kubenswrapper[4559]: I1123 06:47:38.262381 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xk28" event={"ID":"79f79383-e237-4748-bdc0-1edb62550872","Type":"ContainerDied","Data":"6d86471e497eda6109bf5d54d2f2ddb90797a88381ebb9c485fe92531868670f"} Nov 23 06:47:39 crc kubenswrapper[4559]: I1123 06:47:39.271237 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vx7hg" event={"ID":"88092487-aea4-46ae-8918-35e1adf96708","Type":"ContainerStarted","Data":"1c3534e01742e9a5010f933c92e367b76f41b7f36e33872788e59a9faf4f2e2c"} Nov 23 06:47:39 crc kubenswrapper[4559]: I1123 06:47:39.286923 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vx7hg" podStartSLOduration=2.297059738 podStartE2EDuration="21.286909934s" podCreationTimestamp="2025-11-23 06:47:18 +0000 UTC" firstStartedPulling="2025-11-23 06:47:20.077268446 +0000 UTC m=+142.099254060" lastFinishedPulling="2025-11-23 06:47:39.067118643 +0000 UTC m=+161.089104256" observedRunningTime="2025-11-23 06:47:39.282306542 +0000 UTC m=+161.304292156" watchObservedRunningTime="2025-11-23 06:47:39.286909934 +0000 UTC m=+161.308895548" Nov 23 06:47:39 crc kubenswrapper[4559]: I1123 06:47:39.553665 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:47:40 crc kubenswrapper[4559]: I1123 06:47:40.280448 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nbmqd" event={"ID":"7848f717-c7e3-40df-a4d8-854794e3c2e8","Type":"ContainerStarted","Data":"4e2de147f8160b6646d14eddf1cebcd672f7f892812a6cca55a06ee3fe314481"} Nov 23 06:47:40 crc kubenswrapper[4559]: I1123 06:47:40.285848 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m9ltm" event={"ID":"98640d4a-0cbc-428c-8f7a-25a1b36f16f9","Type":"ContainerStarted","Data":"6518fe7506c3af453b8f7ee9cbbe1235ad62a2c03155cb4a853bfc176e155b33"} Nov 23 06:47:40 crc kubenswrapper[4559]: I1123 06:47:40.288054 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xk28" event={"ID":"79f79383-e237-4748-bdc0-1edb62550872","Type":"ContainerStarted","Data":"12518c1c8285c0bd1b138e171fdd4d898e1b2eeb93ebd52e09f5aeea4cdcf338"} Nov 23 06:47:40 crc kubenswrapper[4559]: I1123 06:47:40.289937 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whrv6" event={"ID":"9d1c86c0-8a6f-4914-9593-20ca1220ed8b","Type":"ContainerStarted","Data":"583d1a00ef704e1ed48d88680d0fd97403d80d0a5354bf15dbb8f6eaba7d20e0"} Nov 23 06:47:40 crc kubenswrapper[4559]: I1123 06:47:40.291523 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vqk9p" event={"ID":"25720f84-faae-4d25-8bac-7942852c0a28","Type":"ContainerStarted","Data":"ccc4ab9591897b6a473dcb477f86b18599ad3b75831e10134df0afc89d2f70a0"} Nov 23 06:47:40 crc kubenswrapper[4559]: I1123 06:47:40.293255 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qhvl6" event={"ID":"680dc567-9fb8-4396-adb4-76bb44250c9c","Type":"ContainerStarted","Data":"bf6798357df248f446eddbc97e4fec2a6e2797156dd221864a7f1ef145ef23df"} Nov 23 06:47:40 crc kubenswrapper[4559]: I1123 06:47:40.295604 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rz5lj" event={"ID":"f59cbebb-100b-4d79-83f8-bf1a625caf93","Type":"ContainerStarted","Data":"73b67ebd55b02d6deacf6718d5f6047c50a55e888094a012afe1ccac9b3fdd9e"} Nov 23 06:47:40 crc kubenswrapper[4559]: I1123 06:47:40.296563 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nbmqd" podStartSLOduration=3.275610429 podStartE2EDuration="22.296555281s" podCreationTimestamp="2025-11-23 06:47:18 +0000 UTC" firstStartedPulling="2025-11-23 06:47:20.072253963 +0000 UTC m=+142.094239577" lastFinishedPulling="2025-11-23 06:47:39.093198814 +0000 UTC m=+161.115184429" observedRunningTime="2025-11-23 06:47:40.295832577 +0000 UTC m=+162.317818191" watchObservedRunningTime="2025-11-23 06:47:40.296555281 +0000 UTC m=+162.318540894" Nov 23 06:47:40 crc kubenswrapper[4559]: I1123 06:47:40.311958 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qhvl6" podStartSLOduration=2.285895026 podStartE2EDuration="22.311941396s" podCreationTimestamp="2025-11-23 06:47:18 +0000 UTC" firstStartedPulling="2025-11-23 06:47:19.04716389 +0000 UTC m=+141.069149504" lastFinishedPulling="2025-11-23 06:47:39.07321026 +0000 UTC m=+161.095195874" observedRunningTime="2025-11-23 06:47:40.309820429 +0000 UTC m=+162.331806042" watchObservedRunningTime="2025-11-23 06:47:40.311941396 +0000 UTC m=+162.333927010" Nov 23 06:47:40 crc kubenswrapper[4559]: I1123 06:47:40.327696 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-m9ltm" podStartSLOduration=2.301412275 podStartE2EDuration="20.327681844s" podCreationTimestamp="2025-11-23 06:47:20 +0000 UTC" firstStartedPulling="2025-11-23 06:47:21.083616882 +0000 UTC m=+143.105602496" lastFinishedPulling="2025-11-23 06:47:39.109886452 +0000 UTC m=+161.131872065" observedRunningTime="2025-11-23 06:47:40.327422502 +0000 UTC m=+162.349408117" watchObservedRunningTime="2025-11-23 06:47:40.327681844 +0000 UTC m=+162.349667459" Nov 23 06:47:40 crc kubenswrapper[4559]: I1123 06:47:40.340880 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-whrv6" podStartSLOduration=3.262651963 podStartE2EDuration="22.340865999s" podCreationTimestamp="2025-11-23 06:47:18 +0000 UTC" firstStartedPulling="2025-11-23 06:47:20.066215528 +0000 UTC m=+142.088201142" lastFinishedPulling="2025-11-23 06:47:39.144429564 +0000 UTC m=+161.166415178" observedRunningTime="2025-11-23 06:47:40.33873896 +0000 UTC m=+162.360724574" watchObservedRunningTime="2025-11-23 06:47:40.340865999 +0000 UTC m=+162.362851614" Nov 23 06:47:40 crc kubenswrapper[4559]: I1123 06:47:40.355259 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rz5lj" podStartSLOduration=6.469989947 podStartE2EDuration="19.355246947s" podCreationTimestamp="2025-11-23 06:47:21 +0000 UTC" firstStartedPulling="2025-11-23 06:47:26.192849015 +0000 UTC m=+148.214834629" lastFinishedPulling="2025-11-23 06:47:39.078106015 +0000 UTC m=+161.100091629" observedRunningTime="2025-11-23 06:47:40.353401382 +0000 UTC m=+162.375386996" watchObservedRunningTime="2025-11-23 06:47:40.355246947 +0000 UTC m=+162.377232561" Nov 23 06:47:40 crc kubenswrapper[4559]: I1123 06:47:40.380067 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9xk28" podStartSLOduration=2.345436861 podStartE2EDuration="20.380052048s" podCreationTimestamp="2025-11-23 06:47:20 +0000 UTC" firstStartedPulling="2025-11-23 06:47:21.087179978 +0000 UTC m=+143.109165592" lastFinishedPulling="2025-11-23 06:47:39.121795155 +0000 UTC m=+161.143780779" observedRunningTime="2025-11-23 06:47:40.379302925 +0000 UTC m=+162.401288539" watchObservedRunningTime="2025-11-23 06:47:40.380052048 +0000 UTC m=+162.402037662" Nov 23 06:47:40 crc kubenswrapper[4559]: I1123 06:47:40.390014 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-m9ltm" Nov 23 06:47:40 crc kubenswrapper[4559]: I1123 06:47:40.390048 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-m9ltm" Nov 23 06:47:40 crc kubenswrapper[4559]: I1123 06:47:40.395053 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vqk9p" podStartSLOduration=6.443151404 podStartE2EDuration="19.395040478s" podCreationTimestamp="2025-11-23 06:47:21 +0000 UTC" firstStartedPulling="2025-11-23 06:47:26.180880059 +0000 UTC m=+148.202865673" lastFinishedPulling="2025-11-23 06:47:39.132769133 +0000 UTC m=+161.154754747" observedRunningTime="2025-11-23 06:47:40.392453276 +0000 UTC m=+162.414438890" watchObservedRunningTime="2025-11-23 06:47:40.395040478 +0000 UTC m=+162.417026093" Nov 23 06:47:40 crc kubenswrapper[4559]: I1123 06:47:40.793185 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9xk28" Nov 23 06:47:40 crc kubenswrapper[4559]: I1123 06:47:40.793232 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9xk28" Nov 23 06:47:40 crc kubenswrapper[4559]: I1123 06:47:40.824618 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9xk28" Nov 23 06:47:41 crc kubenswrapper[4559]: I1123 06:47:41.477724 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-m9ltm" podUID="98640d4a-0cbc-428c-8f7a-25a1b36f16f9" containerName="registry-server" probeResult="failure" output=< Nov 23 06:47:41 crc kubenswrapper[4559]: timeout: failed to connect service ":50051" within 1s Nov 23 06:47:41 crc kubenswrapper[4559]: > Nov 23 06:47:41 crc kubenswrapper[4559]: I1123 06:47:41.590309 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rz5lj" Nov 23 06:47:41 crc kubenswrapper[4559]: I1123 06:47:41.590351 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rz5lj" Nov 23 06:47:41 crc kubenswrapper[4559]: I1123 06:47:41.646534 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/abb6827e-709d-4ea9-8090-e4f541b1ff3f-metrics-certs\") pod \"network-metrics-daemon-svb6z\" (UID: \"abb6827e-709d-4ea9-8090-e4f541b1ff3f\") " pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:47:41 crc kubenswrapper[4559]: I1123 06:47:41.652035 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/abb6827e-709d-4ea9-8090-e4f541b1ff3f-metrics-certs\") pod \"network-metrics-daemon-svb6z\" (UID: \"abb6827e-709d-4ea9-8090-e4f541b1ff3f\") " pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:47:41 crc kubenswrapper[4559]: I1123 06:47:41.788987 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-svb6z" Nov 23 06:47:41 crc kubenswrapper[4559]: I1123 06:47:41.990188 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vqk9p" Nov 23 06:47:41 crc kubenswrapper[4559]: I1123 06:47:41.990221 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vqk9p" Nov 23 06:47:42 crc kubenswrapper[4559]: I1123 06:47:42.144694 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-svb6z"] Nov 23 06:47:42 crc kubenswrapper[4559]: W1123 06:47:42.149144 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podabb6827e_709d_4ea9_8090_e4f541b1ff3f.slice/crio-4cca99cb1d0ecc4b16873707cd439b2ab7d1f4022205cb1e9df7579b54b25d86 WatchSource:0}: Error finding container 4cca99cb1d0ecc4b16873707cd439b2ab7d1f4022205cb1e9df7579b54b25d86: Status 404 returned error can't find the container with id 4cca99cb1d0ecc4b16873707cd439b2ab7d1f4022205cb1e9df7579b54b25d86 Nov 23 06:47:42 crc kubenswrapper[4559]: I1123 06:47:42.305276 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-svb6z" event={"ID":"abb6827e-709d-4ea9-8090-e4f541b1ff3f","Type":"ContainerStarted","Data":"4cca99cb1d0ecc4b16873707cd439b2ab7d1f4022205cb1e9df7579b54b25d86"} Nov 23 06:47:42 crc kubenswrapper[4559]: I1123 06:47:42.616873 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rz5lj" podUID="f59cbebb-100b-4d79-83f8-bf1a625caf93" containerName="registry-server" probeResult="failure" output=< Nov 23 06:47:42 crc kubenswrapper[4559]: timeout: failed to connect service ":50051" within 1s Nov 23 06:47:42 crc kubenswrapper[4559]: > Nov 23 06:47:43 crc kubenswrapper[4559]: I1123 06:47:43.017523 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vqk9p" podUID="25720f84-faae-4d25-8bac-7942852c0a28" containerName="registry-server" probeResult="failure" output=< Nov 23 06:47:43 crc kubenswrapper[4559]: timeout: failed to connect service ":50051" within 1s Nov 23 06:47:43 crc kubenswrapper[4559]: > Nov 23 06:47:43 crc kubenswrapper[4559]: I1123 06:47:43.316448 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-svb6z" event={"ID":"abb6827e-709d-4ea9-8090-e4f541b1ff3f","Type":"ContainerStarted","Data":"65768980a68c2563f85a558b85dbb816e0f8bac5633ccd9348ca4aea013a61d6"} Nov 23 06:47:43 crc kubenswrapper[4559]: I1123 06:47:43.316656 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-svb6z" event={"ID":"abb6827e-709d-4ea9-8090-e4f541b1ff3f","Type":"ContainerStarted","Data":"aba583f2c49f0c32f85d2ffccb6291b09bc7592c125b83b7ec0d28337b9e1a87"} Nov 23 06:47:43 crc kubenswrapper[4559]: I1123 06:47:43.740012 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:47:43 crc kubenswrapper[4559]: I1123 06:47:43.743121 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:47:43 crc kubenswrapper[4559]: I1123 06:47:43.770134 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-svb6z" podStartSLOduration=143.770125534 podStartE2EDuration="2m23.770125534s" podCreationTimestamp="2025-11-23 06:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:47:43.330811132 +0000 UTC m=+165.352796746" watchObservedRunningTime="2025-11-23 06:47:43.770125534 +0000 UTC m=+165.792111139" Nov 23 06:47:48 crc kubenswrapper[4559]: I1123 06:47:48.392752 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qhvl6" Nov 23 06:47:48 crc kubenswrapper[4559]: I1123 06:47:48.392960 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qhvl6" Nov 23 06:47:48 crc kubenswrapper[4559]: I1123 06:47:48.420787 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qhvl6" Nov 23 06:47:48 crc kubenswrapper[4559]: I1123 06:47:48.664079 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-whrv6" Nov 23 06:47:48 crc kubenswrapper[4559]: I1123 06:47:48.664118 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-whrv6" Nov 23 06:47:48 crc kubenswrapper[4559]: I1123 06:47:48.691991 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-whrv6" Nov 23 06:47:48 crc kubenswrapper[4559]: I1123 06:47:48.791965 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nbmqd" Nov 23 06:47:48 crc kubenswrapper[4559]: I1123 06:47:48.792003 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nbmqd" Nov 23 06:47:48 crc kubenswrapper[4559]: I1123 06:47:48.818086 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nbmqd" Nov 23 06:47:49 crc kubenswrapper[4559]: I1123 06:47:49.041062 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vx7hg" Nov 23 06:47:49 crc kubenswrapper[4559]: I1123 06:47:49.041332 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vx7hg" Nov 23 06:47:49 crc kubenswrapper[4559]: I1123 06:47:49.066463 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vx7hg" Nov 23 06:47:49 crc kubenswrapper[4559]: I1123 06:47:49.362770 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qhvl6" Nov 23 06:47:49 crc kubenswrapper[4559]: I1123 06:47:49.363037 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-whrv6" Nov 23 06:47:49 crc kubenswrapper[4559]: I1123 06:47:49.363075 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nbmqd" Nov 23 06:47:49 crc kubenswrapper[4559]: I1123 06:47:49.364197 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vx7hg" Nov 23 06:47:50 crc kubenswrapper[4559]: I1123 06:47:50.414804 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-m9ltm" Nov 23 06:47:50 crc kubenswrapper[4559]: I1123 06:47:50.440350 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vx7hg"] Nov 23 06:47:50 crc kubenswrapper[4559]: I1123 06:47:50.444182 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-m9ltm" Nov 23 06:47:50 crc kubenswrapper[4559]: I1123 06:47:50.818424 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9xk28" Nov 23 06:47:51 crc kubenswrapper[4559]: I1123 06:47:51.039613 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nbmqd"] Nov 23 06:47:51 crc kubenswrapper[4559]: I1123 06:47:51.346887 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nbmqd" podUID="7848f717-c7e3-40df-a4d8-854794e3c2e8" containerName="registry-server" containerID="cri-o://4e2de147f8160b6646d14eddf1cebcd672f7f892812a6cca55a06ee3fe314481" gracePeriod=2 Nov 23 06:47:51 crc kubenswrapper[4559]: I1123 06:47:51.617674 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rz5lj" Nov 23 06:47:51 crc kubenswrapper[4559]: I1123 06:47:51.646973 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rz5lj" Nov 23 06:47:51 crc kubenswrapper[4559]: I1123 06:47:51.750606 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nbmqd" Nov 23 06:47:51 crc kubenswrapper[4559]: I1123 06:47:51.840111 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7848f717-c7e3-40df-a4d8-854794e3c2e8-catalog-content\") pod \"7848f717-c7e3-40df-a4d8-854794e3c2e8\" (UID: \"7848f717-c7e3-40df-a4d8-854794e3c2e8\") " Nov 23 06:47:51 crc kubenswrapper[4559]: I1123 06:47:51.840176 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7848f717-c7e3-40df-a4d8-854794e3c2e8-utilities\") pod \"7848f717-c7e3-40df-a4d8-854794e3c2e8\" (UID: \"7848f717-c7e3-40df-a4d8-854794e3c2e8\") " Nov 23 06:47:51 crc kubenswrapper[4559]: I1123 06:47:51.840270 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqxbv\" (UniqueName: \"kubernetes.io/projected/7848f717-c7e3-40df-a4d8-854794e3c2e8-kube-api-access-dqxbv\") pod \"7848f717-c7e3-40df-a4d8-854794e3c2e8\" (UID: \"7848f717-c7e3-40df-a4d8-854794e3c2e8\") " Nov 23 06:47:51 crc kubenswrapper[4559]: I1123 06:47:51.840792 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7848f717-c7e3-40df-a4d8-854794e3c2e8-utilities" (OuterVolumeSpecName: "utilities") pod "7848f717-c7e3-40df-a4d8-854794e3c2e8" (UID: "7848f717-c7e3-40df-a4d8-854794e3c2e8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:47:51 crc kubenswrapper[4559]: I1123 06:47:51.844428 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7848f717-c7e3-40df-a4d8-854794e3c2e8-kube-api-access-dqxbv" (OuterVolumeSpecName: "kube-api-access-dqxbv") pod "7848f717-c7e3-40df-a4d8-854794e3c2e8" (UID: "7848f717-c7e3-40df-a4d8-854794e3c2e8"). InnerVolumeSpecName "kube-api-access-dqxbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:47:51 crc kubenswrapper[4559]: I1123 06:47:51.872106 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7848f717-c7e3-40df-a4d8-854794e3c2e8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7848f717-c7e3-40df-a4d8-854794e3c2e8" (UID: "7848f717-c7e3-40df-a4d8-854794e3c2e8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:47:51 crc kubenswrapper[4559]: I1123 06:47:51.941864 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqxbv\" (UniqueName: \"kubernetes.io/projected/7848f717-c7e3-40df-a4d8-854794e3c2e8-kube-api-access-dqxbv\") on node \"crc\" DevicePath \"\"" Nov 23 06:47:51 crc kubenswrapper[4559]: I1123 06:47:51.941896 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7848f717-c7e3-40df-a4d8-854794e3c2e8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 06:47:51 crc kubenswrapper[4559]: I1123 06:47:51.941906 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7848f717-c7e3-40df-a4d8-854794e3c2e8-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.014431 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vqk9p" Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.038753 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vqk9p" Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.351758 4559 generic.go:334] "Generic (PLEG): container finished" podID="7848f717-c7e3-40df-a4d8-854794e3c2e8" containerID="4e2de147f8160b6646d14eddf1cebcd672f7f892812a6cca55a06ee3fe314481" exitCode=0 Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.351795 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nbmqd" Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.351842 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nbmqd" event={"ID":"7848f717-c7e3-40df-a4d8-854794e3c2e8","Type":"ContainerDied","Data":"4e2de147f8160b6646d14eddf1cebcd672f7f892812a6cca55a06ee3fe314481"} Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.351884 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nbmqd" event={"ID":"7848f717-c7e3-40df-a4d8-854794e3c2e8","Type":"ContainerDied","Data":"6959e76b09d0f5832cf28d8f4e6c10307973653e13210052a5488c75d73fae85"} Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.351907 4559 scope.go:117] "RemoveContainer" containerID="4e2de147f8160b6646d14eddf1cebcd672f7f892812a6cca55a06ee3fe314481" Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.352136 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vx7hg" podUID="88092487-aea4-46ae-8918-35e1adf96708" containerName="registry-server" containerID="cri-o://1c3534e01742e9a5010f933c92e367b76f41b7f36e33872788e59a9faf4f2e2c" gracePeriod=2 Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.365570 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nbmqd"] Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.367211 4559 scope.go:117] "RemoveContainer" containerID="ab1ed9a1bf45ad6d3d3347dff288a670fcf80fbcf3b24cd1b0931a98faeec8d6" Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.367442 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nbmqd"] Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.380494 4559 scope.go:117] "RemoveContainer" containerID="bc6b991c3573a01c8577e700208bc288b17ebe5a7dae751940e70f16219a0aa7" Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.436273 4559 scope.go:117] "RemoveContainer" containerID="4e2de147f8160b6646d14eddf1cebcd672f7f892812a6cca55a06ee3fe314481" Nov 23 06:47:52 crc kubenswrapper[4559]: E1123 06:47:52.436580 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e2de147f8160b6646d14eddf1cebcd672f7f892812a6cca55a06ee3fe314481\": container with ID starting with 4e2de147f8160b6646d14eddf1cebcd672f7f892812a6cca55a06ee3fe314481 not found: ID does not exist" containerID="4e2de147f8160b6646d14eddf1cebcd672f7f892812a6cca55a06ee3fe314481" Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.436613 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e2de147f8160b6646d14eddf1cebcd672f7f892812a6cca55a06ee3fe314481"} err="failed to get container status \"4e2de147f8160b6646d14eddf1cebcd672f7f892812a6cca55a06ee3fe314481\": rpc error: code = NotFound desc = could not find container \"4e2de147f8160b6646d14eddf1cebcd672f7f892812a6cca55a06ee3fe314481\": container with ID starting with 4e2de147f8160b6646d14eddf1cebcd672f7f892812a6cca55a06ee3fe314481 not found: ID does not exist" Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.436789 4559 scope.go:117] "RemoveContainer" containerID="ab1ed9a1bf45ad6d3d3347dff288a670fcf80fbcf3b24cd1b0931a98faeec8d6" Nov 23 06:47:52 crc kubenswrapper[4559]: E1123 06:47:52.437149 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab1ed9a1bf45ad6d3d3347dff288a670fcf80fbcf3b24cd1b0931a98faeec8d6\": container with ID starting with ab1ed9a1bf45ad6d3d3347dff288a670fcf80fbcf3b24cd1b0931a98faeec8d6 not found: ID does not exist" containerID="ab1ed9a1bf45ad6d3d3347dff288a670fcf80fbcf3b24cd1b0931a98faeec8d6" Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.437170 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab1ed9a1bf45ad6d3d3347dff288a670fcf80fbcf3b24cd1b0931a98faeec8d6"} err="failed to get container status \"ab1ed9a1bf45ad6d3d3347dff288a670fcf80fbcf3b24cd1b0931a98faeec8d6\": rpc error: code = NotFound desc = could not find container \"ab1ed9a1bf45ad6d3d3347dff288a670fcf80fbcf3b24cd1b0931a98faeec8d6\": container with ID starting with ab1ed9a1bf45ad6d3d3347dff288a670fcf80fbcf3b24cd1b0931a98faeec8d6 not found: ID does not exist" Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.437189 4559 scope.go:117] "RemoveContainer" containerID="bc6b991c3573a01c8577e700208bc288b17ebe5a7dae751940e70f16219a0aa7" Nov 23 06:47:52 crc kubenswrapper[4559]: E1123 06:47:52.437440 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc6b991c3573a01c8577e700208bc288b17ebe5a7dae751940e70f16219a0aa7\": container with ID starting with bc6b991c3573a01c8577e700208bc288b17ebe5a7dae751940e70f16219a0aa7 not found: ID does not exist" containerID="bc6b991c3573a01c8577e700208bc288b17ebe5a7dae751940e70f16219a0aa7" Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.437468 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc6b991c3573a01c8577e700208bc288b17ebe5a7dae751940e70f16219a0aa7"} err="failed to get container status \"bc6b991c3573a01c8577e700208bc288b17ebe5a7dae751940e70f16219a0aa7\": rpc error: code = NotFound desc = could not find container \"bc6b991c3573a01c8577e700208bc288b17ebe5a7dae751940e70f16219a0aa7\": container with ID starting with bc6b991c3573a01c8577e700208bc288b17ebe5a7dae751940e70f16219a0aa7 not found: ID does not exist" Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.731477 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vx7hg" Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.842435 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9xk28"] Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.842718 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9xk28" podUID="79f79383-e237-4748-bdc0-1edb62550872" containerName="registry-server" containerID="cri-o://12518c1c8285c0bd1b138e171fdd4d898e1b2eeb93ebd52e09f5aeea4cdcf338" gracePeriod=2 Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.851121 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kq8sm\" (UniqueName: \"kubernetes.io/projected/88092487-aea4-46ae-8918-35e1adf96708-kube-api-access-kq8sm\") pod \"88092487-aea4-46ae-8918-35e1adf96708\" (UID: \"88092487-aea4-46ae-8918-35e1adf96708\") " Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.851209 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88092487-aea4-46ae-8918-35e1adf96708-utilities\") pod \"88092487-aea4-46ae-8918-35e1adf96708\" (UID: \"88092487-aea4-46ae-8918-35e1adf96708\") " Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.851257 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88092487-aea4-46ae-8918-35e1adf96708-catalog-content\") pod \"88092487-aea4-46ae-8918-35e1adf96708\" (UID: \"88092487-aea4-46ae-8918-35e1adf96708\") " Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.851865 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88092487-aea4-46ae-8918-35e1adf96708-utilities" (OuterVolumeSpecName: "utilities") pod "88092487-aea4-46ae-8918-35e1adf96708" (UID: "88092487-aea4-46ae-8918-35e1adf96708"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.853530 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88092487-aea4-46ae-8918-35e1adf96708-kube-api-access-kq8sm" (OuterVolumeSpecName: "kube-api-access-kq8sm") pod "88092487-aea4-46ae-8918-35e1adf96708" (UID: "88092487-aea4-46ae-8918-35e1adf96708"). InnerVolumeSpecName "kube-api-access-kq8sm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.888032 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88092487-aea4-46ae-8918-35e1adf96708-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "88092487-aea4-46ae-8918-35e1adf96708" (UID: "88092487-aea4-46ae-8918-35e1adf96708"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.952424 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kq8sm\" (UniqueName: \"kubernetes.io/projected/88092487-aea4-46ae-8918-35e1adf96708-kube-api-access-kq8sm\") on node \"crc\" DevicePath \"\"" Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.952458 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88092487-aea4-46ae-8918-35e1adf96708-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 06:47:52 crc kubenswrapper[4559]: I1123 06:47:52.952468 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88092487-aea4-46ae-8918-35e1adf96708-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.182461 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9xk28" Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.255631 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79f79383-e237-4748-bdc0-1edb62550872-catalog-content\") pod \"79f79383-e237-4748-bdc0-1edb62550872\" (UID: \"79f79383-e237-4748-bdc0-1edb62550872\") " Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.255772 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9slv6\" (UniqueName: \"kubernetes.io/projected/79f79383-e237-4748-bdc0-1edb62550872-kube-api-access-9slv6\") pod \"79f79383-e237-4748-bdc0-1edb62550872\" (UID: \"79f79383-e237-4748-bdc0-1edb62550872\") " Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.255801 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79f79383-e237-4748-bdc0-1edb62550872-utilities\") pod \"79f79383-e237-4748-bdc0-1edb62550872\" (UID: \"79f79383-e237-4748-bdc0-1edb62550872\") " Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.256484 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79f79383-e237-4748-bdc0-1edb62550872-utilities" (OuterVolumeSpecName: "utilities") pod "79f79383-e237-4748-bdc0-1edb62550872" (UID: "79f79383-e237-4748-bdc0-1edb62550872"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.258404 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79f79383-e237-4748-bdc0-1edb62550872-kube-api-access-9slv6" (OuterVolumeSpecName: "kube-api-access-9slv6") pod "79f79383-e237-4748-bdc0-1edb62550872" (UID: "79f79383-e237-4748-bdc0-1edb62550872"). InnerVolumeSpecName "kube-api-access-9slv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.268098 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79f79383-e237-4748-bdc0-1edb62550872-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "79f79383-e237-4748-bdc0-1edb62550872" (UID: "79f79383-e237-4748-bdc0-1edb62550872"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.356513 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79f79383-e237-4748-bdc0-1edb62550872-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.356624 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9slv6\" (UniqueName: \"kubernetes.io/projected/79f79383-e237-4748-bdc0-1edb62550872-kube-api-access-9slv6\") on node \"crc\" DevicePath \"\"" Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.356719 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79f79383-e237-4748-bdc0-1edb62550872-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.358873 4559 generic.go:334] "Generic (PLEG): container finished" podID="79f79383-e237-4748-bdc0-1edb62550872" containerID="12518c1c8285c0bd1b138e171fdd4d898e1b2eeb93ebd52e09f5aeea4cdcf338" exitCode=0 Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.358918 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xk28" event={"ID":"79f79383-e237-4748-bdc0-1edb62550872","Type":"ContainerDied","Data":"12518c1c8285c0bd1b138e171fdd4d898e1b2eeb93ebd52e09f5aeea4cdcf338"} Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.358938 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xk28" event={"ID":"79f79383-e237-4748-bdc0-1edb62550872","Type":"ContainerDied","Data":"3f705a6d15152768eb7af2a5502224b0f6e9ac81fcc9e0ec454f6b8a7dec31a4"} Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.358952 4559 scope.go:117] "RemoveContainer" containerID="12518c1c8285c0bd1b138e171fdd4d898e1b2eeb93ebd52e09f5aeea4cdcf338" Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.359035 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9xk28" Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.363608 4559 generic.go:334] "Generic (PLEG): container finished" podID="88092487-aea4-46ae-8918-35e1adf96708" containerID="1c3534e01742e9a5010f933c92e367b76f41b7f36e33872788e59a9faf4f2e2c" exitCode=0 Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.363630 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vx7hg" event={"ID":"88092487-aea4-46ae-8918-35e1adf96708","Type":"ContainerDied","Data":"1c3534e01742e9a5010f933c92e367b76f41b7f36e33872788e59a9faf4f2e2c"} Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.363688 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vx7hg" event={"ID":"88092487-aea4-46ae-8918-35e1adf96708","Type":"ContainerDied","Data":"7f2614485ae243f6229a0ccd0452f92235464eec30867b99d75d6cf9195d97a9"} Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.363694 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vx7hg" Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.373554 4559 scope.go:117] "RemoveContainer" containerID="6d86471e497eda6109bf5d54d2f2ddb90797a88381ebb9c485fe92531868670f" Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.379960 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9xk28"] Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.382979 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9xk28"] Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.390972 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vx7hg"] Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.393250 4559 scope.go:117] "RemoveContainer" containerID="1b4be2cec437fdc1bc4a6614ef3fe251bb7a7c8e410dd4f477165e2778c74f08" Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.393968 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vx7hg"] Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.418152 4559 scope.go:117] "RemoveContainer" containerID="12518c1c8285c0bd1b138e171fdd4d898e1b2eeb93ebd52e09f5aeea4cdcf338" Nov 23 06:47:53 crc kubenswrapper[4559]: E1123 06:47:53.418540 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12518c1c8285c0bd1b138e171fdd4d898e1b2eeb93ebd52e09f5aeea4cdcf338\": container with ID starting with 12518c1c8285c0bd1b138e171fdd4d898e1b2eeb93ebd52e09f5aeea4cdcf338 not found: ID does not exist" containerID="12518c1c8285c0bd1b138e171fdd4d898e1b2eeb93ebd52e09f5aeea4cdcf338" Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.418608 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12518c1c8285c0bd1b138e171fdd4d898e1b2eeb93ebd52e09f5aeea4cdcf338"} err="failed to get container status \"12518c1c8285c0bd1b138e171fdd4d898e1b2eeb93ebd52e09f5aeea4cdcf338\": rpc error: code = NotFound desc = could not find container \"12518c1c8285c0bd1b138e171fdd4d898e1b2eeb93ebd52e09f5aeea4cdcf338\": container with ID starting with 12518c1c8285c0bd1b138e171fdd4d898e1b2eeb93ebd52e09f5aeea4cdcf338 not found: ID does not exist" Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.418627 4559 scope.go:117] "RemoveContainer" containerID="6d86471e497eda6109bf5d54d2f2ddb90797a88381ebb9c485fe92531868670f" Nov 23 06:47:53 crc kubenswrapper[4559]: E1123 06:47:53.418968 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d86471e497eda6109bf5d54d2f2ddb90797a88381ebb9c485fe92531868670f\": container with ID starting with 6d86471e497eda6109bf5d54d2f2ddb90797a88381ebb9c485fe92531868670f not found: ID does not exist" containerID="6d86471e497eda6109bf5d54d2f2ddb90797a88381ebb9c485fe92531868670f" Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.419011 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d86471e497eda6109bf5d54d2f2ddb90797a88381ebb9c485fe92531868670f"} err="failed to get container status \"6d86471e497eda6109bf5d54d2f2ddb90797a88381ebb9c485fe92531868670f\": rpc error: code = NotFound desc = could not find container \"6d86471e497eda6109bf5d54d2f2ddb90797a88381ebb9c485fe92531868670f\": container with ID starting with 6d86471e497eda6109bf5d54d2f2ddb90797a88381ebb9c485fe92531868670f not found: ID does not exist" Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.419031 4559 scope.go:117] "RemoveContainer" containerID="1b4be2cec437fdc1bc4a6614ef3fe251bb7a7c8e410dd4f477165e2778c74f08" Nov 23 06:47:53 crc kubenswrapper[4559]: E1123 06:47:53.419296 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b4be2cec437fdc1bc4a6614ef3fe251bb7a7c8e410dd4f477165e2778c74f08\": container with ID starting with 1b4be2cec437fdc1bc4a6614ef3fe251bb7a7c8e410dd4f477165e2778c74f08 not found: ID does not exist" containerID="1b4be2cec437fdc1bc4a6614ef3fe251bb7a7c8e410dd4f477165e2778c74f08" Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.419376 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b4be2cec437fdc1bc4a6614ef3fe251bb7a7c8e410dd4f477165e2778c74f08"} err="failed to get container status \"1b4be2cec437fdc1bc4a6614ef3fe251bb7a7c8e410dd4f477165e2778c74f08\": rpc error: code = NotFound desc = could not find container \"1b4be2cec437fdc1bc4a6614ef3fe251bb7a7c8e410dd4f477165e2778c74f08\": container with ID starting with 1b4be2cec437fdc1bc4a6614ef3fe251bb7a7c8e410dd4f477165e2778c74f08 not found: ID does not exist" Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.419439 4559 scope.go:117] "RemoveContainer" containerID="1c3534e01742e9a5010f933c92e367b76f41b7f36e33872788e59a9faf4f2e2c" Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.429426 4559 scope.go:117] "RemoveContainer" containerID="196d20033d13965be63bc2490bdb1a7109970eb0f96f0ebc2195ce983ae494da" Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.441900 4559 scope.go:117] "RemoveContainer" containerID="09e85832676ae7db5010b3d73350a4c2b5589ad6bcabeaf3a7db0494ca356285" Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.452047 4559 scope.go:117] "RemoveContainer" containerID="1c3534e01742e9a5010f933c92e367b76f41b7f36e33872788e59a9faf4f2e2c" Nov 23 06:47:53 crc kubenswrapper[4559]: E1123 06:47:53.452287 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c3534e01742e9a5010f933c92e367b76f41b7f36e33872788e59a9faf4f2e2c\": container with ID starting with 1c3534e01742e9a5010f933c92e367b76f41b7f36e33872788e59a9faf4f2e2c not found: ID does not exist" containerID="1c3534e01742e9a5010f933c92e367b76f41b7f36e33872788e59a9faf4f2e2c" Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.452314 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c3534e01742e9a5010f933c92e367b76f41b7f36e33872788e59a9faf4f2e2c"} err="failed to get container status \"1c3534e01742e9a5010f933c92e367b76f41b7f36e33872788e59a9faf4f2e2c\": rpc error: code = NotFound desc = could not find container \"1c3534e01742e9a5010f933c92e367b76f41b7f36e33872788e59a9faf4f2e2c\": container with ID starting with 1c3534e01742e9a5010f933c92e367b76f41b7f36e33872788e59a9faf4f2e2c not found: ID does not exist" Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.452328 4559 scope.go:117] "RemoveContainer" containerID="196d20033d13965be63bc2490bdb1a7109970eb0f96f0ebc2195ce983ae494da" Nov 23 06:47:53 crc kubenswrapper[4559]: E1123 06:47:53.452553 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"196d20033d13965be63bc2490bdb1a7109970eb0f96f0ebc2195ce983ae494da\": container with ID starting with 196d20033d13965be63bc2490bdb1a7109970eb0f96f0ebc2195ce983ae494da not found: ID does not exist" containerID="196d20033d13965be63bc2490bdb1a7109970eb0f96f0ebc2195ce983ae494da" Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.452580 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"196d20033d13965be63bc2490bdb1a7109970eb0f96f0ebc2195ce983ae494da"} err="failed to get container status \"196d20033d13965be63bc2490bdb1a7109970eb0f96f0ebc2195ce983ae494da\": rpc error: code = NotFound desc = could not find container \"196d20033d13965be63bc2490bdb1a7109970eb0f96f0ebc2195ce983ae494da\": container with ID starting with 196d20033d13965be63bc2490bdb1a7109970eb0f96f0ebc2195ce983ae494da not found: ID does not exist" Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.452607 4559 scope.go:117] "RemoveContainer" containerID="09e85832676ae7db5010b3d73350a4c2b5589ad6bcabeaf3a7db0494ca356285" Nov 23 06:47:53 crc kubenswrapper[4559]: E1123 06:47:53.452805 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09e85832676ae7db5010b3d73350a4c2b5589ad6bcabeaf3a7db0494ca356285\": container with ID starting with 09e85832676ae7db5010b3d73350a4c2b5589ad6bcabeaf3a7db0494ca356285 not found: ID does not exist" containerID="09e85832676ae7db5010b3d73350a4c2b5589ad6bcabeaf3a7db0494ca356285" Nov 23 06:47:53 crc kubenswrapper[4559]: I1123 06:47:53.452825 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09e85832676ae7db5010b3d73350a4c2b5589ad6bcabeaf3a7db0494ca356285"} err="failed to get container status \"09e85832676ae7db5010b3d73350a4c2b5589ad6bcabeaf3a7db0494ca356285\": rpc error: code = NotFound desc = could not find container \"09e85832676ae7db5010b3d73350a4c2b5589ad6bcabeaf3a7db0494ca356285\": container with ID starting with 09e85832676ae7db5010b3d73350a4c2b5589ad6bcabeaf3a7db0494ca356285 not found: ID does not exist" Nov 23 06:47:54 crc kubenswrapper[4559]: I1123 06:47:54.154081 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-r779q" Nov 23 06:47:54 crc kubenswrapper[4559]: I1123 06:47:54.278779 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7848f717-c7e3-40df-a4d8-854794e3c2e8" path="/var/lib/kubelet/pods/7848f717-c7e3-40df-a4d8-854794e3c2e8/volumes" Nov 23 06:47:54 crc kubenswrapper[4559]: I1123 06:47:54.279316 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79f79383-e237-4748-bdc0-1edb62550872" path="/var/lib/kubelet/pods/79f79383-e237-4748-bdc0-1edb62550872/volumes" Nov 23 06:47:54 crc kubenswrapper[4559]: I1123 06:47:54.279842 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88092487-aea4-46ae-8918-35e1adf96708" path="/var/lib/kubelet/pods/88092487-aea4-46ae-8918-35e1adf96708/volumes" Nov 23 06:47:55 crc kubenswrapper[4559]: I1123 06:47:55.438377 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vqk9p"] Nov 23 06:47:55 crc kubenswrapper[4559]: I1123 06:47:55.438748 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vqk9p" podUID="25720f84-faae-4d25-8bac-7942852c0a28" containerName="registry-server" containerID="cri-o://ccc4ab9591897b6a473dcb477f86b18599ad3b75831e10134df0afc89d2f70a0" gracePeriod=2 Nov 23 06:47:55 crc kubenswrapper[4559]: I1123 06:47:55.790601 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vqk9p" Nov 23 06:47:55 crc kubenswrapper[4559]: I1123 06:47:55.879023 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25720f84-faae-4d25-8bac-7942852c0a28-catalog-content\") pod \"25720f84-faae-4d25-8bac-7942852c0a28\" (UID: \"25720f84-faae-4d25-8bac-7942852c0a28\") " Nov 23 06:47:55 crc kubenswrapper[4559]: I1123 06:47:55.879090 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5psk\" (UniqueName: \"kubernetes.io/projected/25720f84-faae-4d25-8bac-7942852c0a28-kube-api-access-c5psk\") pod \"25720f84-faae-4d25-8bac-7942852c0a28\" (UID: \"25720f84-faae-4d25-8bac-7942852c0a28\") " Nov 23 06:47:55 crc kubenswrapper[4559]: I1123 06:47:55.879131 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25720f84-faae-4d25-8bac-7942852c0a28-utilities\") pod \"25720f84-faae-4d25-8bac-7942852c0a28\" (UID: \"25720f84-faae-4d25-8bac-7942852c0a28\") " Nov 23 06:47:55 crc kubenswrapper[4559]: I1123 06:47:55.879771 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25720f84-faae-4d25-8bac-7942852c0a28-utilities" (OuterVolumeSpecName: "utilities") pod "25720f84-faae-4d25-8bac-7942852c0a28" (UID: "25720f84-faae-4d25-8bac-7942852c0a28"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:47:55 crc kubenswrapper[4559]: I1123 06:47:55.882819 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25720f84-faae-4d25-8bac-7942852c0a28-kube-api-access-c5psk" (OuterVolumeSpecName: "kube-api-access-c5psk") pod "25720f84-faae-4d25-8bac-7942852c0a28" (UID: "25720f84-faae-4d25-8bac-7942852c0a28"). InnerVolumeSpecName "kube-api-access-c5psk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:47:55 crc kubenswrapper[4559]: I1123 06:47:55.937807 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25720f84-faae-4d25-8bac-7942852c0a28-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "25720f84-faae-4d25-8bac-7942852c0a28" (UID: "25720f84-faae-4d25-8bac-7942852c0a28"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:47:55 crc kubenswrapper[4559]: I1123 06:47:55.979979 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25720f84-faae-4d25-8bac-7942852c0a28-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 06:47:55 crc kubenswrapper[4559]: I1123 06:47:55.980004 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5psk\" (UniqueName: \"kubernetes.io/projected/25720f84-faae-4d25-8bac-7942852c0a28-kube-api-access-c5psk\") on node \"crc\" DevicePath \"\"" Nov 23 06:47:55 crc kubenswrapper[4559]: I1123 06:47:55.980017 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25720f84-faae-4d25-8bac-7942852c0a28-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 06:47:56 crc kubenswrapper[4559]: I1123 06:47:56.166826 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 06:47:56 crc kubenswrapper[4559]: I1123 06:47:56.166867 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 06:47:56 crc kubenswrapper[4559]: I1123 06:47:56.377137 4559 generic.go:334] "Generic (PLEG): container finished" podID="25720f84-faae-4d25-8bac-7942852c0a28" containerID="ccc4ab9591897b6a473dcb477f86b18599ad3b75831e10134df0afc89d2f70a0" exitCode=0 Nov 23 06:47:56 crc kubenswrapper[4559]: I1123 06:47:56.377175 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vqk9p" event={"ID":"25720f84-faae-4d25-8bac-7942852c0a28","Type":"ContainerDied","Data":"ccc4ab9591897b6a473dcb477f86b18599ad3b75831e10134df0afc89d2f70a0"} Nov 23 06:47:56 crc kubenswrapper[4559]: I1123 06:47:56.377200 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vqk9p" event={"ID":"25720f84-faae-4d25-8bac-7942852c0a28","Type":"ContainerDied","Data":"3a78d83385f885c7edaab01d9bfa79634ee9de09d9874ba478617d73e8cbe073"} Nov 23 06:47:56 crc kubenswrapper[4559]: I1123 06:47:56.377200 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vqk9p" Nov 23 06:47:56 crc kubenswrapper[4559]: I1123 06:47:56.377219 4559 scope.go:117] "RemoveContainer" containerID="ccc4ab9591897b6a473dcb477f86b18599ad3b75831e10134df0afc89d2f70a0" Nov 23 06:47:56 crc kubenswrapper[4559]: I1123 06:47:56.389484 4559 scope.go:117] "RemoveContainer" containerID="c9b71830cd1aa9010ffd3021cff2c654949bf62d94c1757afad72aeae26d025f" Nov 23 06:47:56 crc kubenswrapper[4559]: I1123 06:47:56.389665 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vqk9p"] Nov 23 06:47:56 crc kubenswrapper[4559]: I1123 06:47:56.392060 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vqk9p"] Nov 23 06:47:56 crc kubenswrapper[4559]: I1123 06:47:56.401178 4559 scope.go:117] "RemoveContainer" containerID="b57b779dbe3d6ca04aac0f593a45a01c1098b6c056c2e1fe1f91cd5e5bb53073" Nov 23 06:47:56 crc kubenswrapper[4559]: I1123 06:47:56.411791 4559 scope.go:117] "RemoveContainer" containerID="ccc4ab9591897b6a473dcb477f86b18599ad3b75831e10134df0afc89d2f70a0" Nov 23 06:47:56 crc kubenswrapper[4559]: E1123 06:47:56.412186 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccc4ab9591897b6a473dcb477f86b18599ad3b75831e10134df0afc89d2f70a0\": container with ID starting with ccc4ab9591897b6a473dcb477f86b18599ad3b75831e10134df0afc89d2f70a0 not found: ID does not exist" containerID="ccc4ab9591897b6a473dcb477f86b18599ad3b75831e10134df0afc89d2f70a0" Nov 23 06:47:56 crc kubenswrapper[4559]: I1123 06:47:56.412222 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccc4ab9591897b6a473dcb477f86b18599ad3b75831e10134df0afc89d2f70a0"} err="failed to get container status \"ccc4ab9591897b6a473dcb477f86b18599ad3b75831e10134df0afc89d2f70a0\": rpc error: code = NotFound desc = could not find container \"ccc4ab9591897b6a473dcb477f86b18599ad3b75831e10134df0afc89d2f70a0\": container with ID starting with ccc4ab9591897b6a473dcb477f86b18599ad3b75831e10134df0afc89d2f70a0 not found: ID does not exist" Nov 23 06:47:56 crc kubenswrapper[4559]: I1123 06:47:56.412248 4559 scope.go:117] "RemoveContainer" containerID="c9b71830cd1aa9010ffd3021cff2c654949bf62d94c1757afad72aeae26d025f" Nov 23 06:47:56 crc kubenswrapper[4559]: E1123 06:47:56.412467 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9b71830cd1aa9010ffd3021cff2c654949bf62d94c1757afad72aeae26d025f\": container with ID starting with c9b71830cd1aa9010ffd3021cff2c654949bf62d94c1757afad72aeae26d025f not found: ID does not exist" containerID="c9b71830cd1aa9010ffd3021cff2c654949bf62d94c1757afad72aeae26d025f" Nov 23 06:47:56 crc kubenswrapper[4559]: I1123 06:47:56.412495 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9b71830cd1aa9010ffd3021cff2c654949bf62d94c1757afad72aeae26d025f"} err="failed to get container status \"c9b71830cd1aa9010ffd3021cff2c654949bf62d94c1757afad72aeae26d025f\": rpc error: code = NotFound desc = could not find container \"c9b71830cd1aa9010ffd3021cff2c654949bf62d94c1757afad72aeae26d025f\": container with ID starting with c9b71830cd1aa9010ffd3021cff2c654949bf62d94c1757afad72aeae26d025f not found: ID does not exist" Nov 23 06:47:56 crc kubenswrapper[4559]: I1123 06:47:56.412516 4559 scope.go:117] "RemoveContainer" containerID="b57b779dbe3d6ca04aac0f593a45a01c1098b6c056c2e1fe1f91cd5e5bb53073" Nov 23 06:47:56 crc kubenswrapper[4559]: E1123 06:47:56.412933 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b57b779dbe3d6ca04aac0f593a45a01c1098b6c056c2e1fe1f91cd5e5bb53073\": container with ID starting with b57b779dbe3d6ca04aac0f593a45a01c1098b6c056c2e1fe1f91cd5e5bb53073 not found: ID does not exist" containerID="b57b779dbe3d6ca04aac0f593a45a01c1098b6c056c2e1fe1f91cd5e5bb53073" Nov 23 06:47:56 crc kubenswrapper[4559]: I1123 06:47:56.412965 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b57b779dbe3d6ca04aac0f593a45a01c1098b6c056c2e1fe1f91cd5e5bb53073"} err="failed to get container status \"b57b779dbe3d6ca04aac0f593a45a01c1098b6c056c2e1fe1f91cd5e5bb53073\": rpc error: code = NotFound desc = could not find container \"b57b779dbe3d6ca04aac0f593a45a01c1098b6c056c2e1fe1f91cd5e5bb53073\": container with ID starting with b57b779dbe3d6ca04aac0f593a45a01c1098b6c056c2e1fe1f91cd5e5bb53073 not found: ID does not exist" Nov 23 06:47:58 crc kubenswrapper[4559]: I1123 06:47:58.281379 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25720f84-faae-4d25-8bac-7942852c0a28" path="/var/lib/kubelet/pods/25720f84-faae-4d25-8bac-7942852c0a28/volumes" Nov 23 06:48:04 crc kubenswrapper[4559]: I1123 06:48:04.293434 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 06:48:12 crc kubenswrapper[4559]: I1123 06:48:12.932909 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-kg77n"] Nov 23 06:48:26 crc kubenswrapper[4559]: I1123 06:48:26.167276 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 06:48:26 crc kubenswrapper[4559]: I1123 06:48:26.167708 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 06:48:26 crc kubenswrapper[4559]: I1123 06:48:26.167762 4559 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" Nov 23 06:48:26 crc kubenswrapper[4559]: I1123 06:48:26.168186 4559 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43"} pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 06:48:26 crc kubenswrapper[4559]: I1123 06:48:26.168239 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" containerID="cri-o://053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43" gracePeriod=600 Nov 23 06:48:26 crc kubenswrapper[4559]: I1123 06:48:26.482940 4559 generic.go:334] "Generic (PLEG): container finished" podID="4731beee-0cac-4189-8a70-743b0b709095" containerID="053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43" exitCode=0 Nov 23 06:48:26 crc kubenswrapper[4559]: I1123 06:48:26.483013 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerDied","Data":"053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43"} Nov 23 06:48:26 crc kubenswrapper[4559]: I1123 06:48:26.483137 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerStarted","Data":"232a78e72da23760d36189f8c69ea7f78803b94ef0a735cc1b9c323800c5683a"} Nov 23 06:48:37 crc kubenswrapper[4559]: I1123 06:48:37.951054 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" podUID="e7b5892a-8920-4cef-94a1-3736328e9a39" containerName="oauth-openshift" containerID="cri-o://ce69456f4d4198e8b9edc01bd976a0de7270cf5baf44e52b38599f01a9ab6c82" gracePeriod=15 Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.229763 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.250864 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7c47d65989-g8qpb"] Nov 23 06:48:38 crc kubenswrapper[4559]: E1123 06:48:38.251035 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7848f717-c7e3-40df-a4d8-854794e3c2e8" containerName="extract-utilities" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.251052 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="7848f717-c7e3-40df-a4d8-854794e3c2e8" containerName="extract-utilities" Nov 23 06:48:38 crc kubenswrapper[4559]: E1123 06:48:38.251061 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88092487-aea4-46ae-8918-35e1adf96708" containerName="extract-content" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.251067 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="88092487-aea4-46ae-8918-35e1adf96708" containerName="extract-content" Nov 23 06:48:38 crc kubenswrapper[4559]: E1123 06:48:38.251077 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88092487-aea4-46ae-8918-35e1adf96708" containerName="extract-utilities" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.251083 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="88092487-aea4-46ae-8918-35e1adf96708" containerName="extract-utilities" Nov 23 06:48:38 crc kubenswrapper[4559]: E1123 06:48:38.251090 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eccd095a-81da-4c51-8107-0ad84257c648" containerName="pruner" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.251096 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="eccd095a-81da-4c51-8107-0ad84257c648" containerName="pruner" Nov 23 06:48:38 crc kubenswrapper[4559]: E1123 06:48:38.251102 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7848f717-c7e3-40df-a4d8-854794e3c2e8" containerName="registry-server" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.251109 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="7848f717-c7e3-40df-a4d8-854794e3c2e8" containerName="registry-server" Nov 23 06:48:38 crc kubenswrapper[4559]: E1123 06:48:38.251118 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88092487-aea4-46ae-8918-35e1adf96708" containerName="registry-server" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.251124 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="88092487-aea4-46ae-8918-35e1adf96708" containerName="registry-server" Nov 23 06:48:38 crc kubenswrapper[4559]: E1123 06:48:38.251132 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7848f717-c7e3-40df-a4d8-854794e3c2e8" containerName="extract-content" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.251137 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="7848f717-c7e3-40df-a4d8-854794e3c2e8" containerName="extract-content" Nov 23 06:48:38 crc kubenswrapper[4559]: E1123 06:48:38.251144 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25720f84-faae-4d25-8bac-7942852c0a28" containerName="registry-server" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.251149 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="25720f84-faae-4d25-8bac-7942852c0a28" containerName="registry-server" Nov 23 06:48:38 crc kubenswrapper[4559]: E1123 06:48:38.251156 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21b0a5f3-85fe-4f27-a139-3f1d21b063fa" containerName="pruner" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.251161 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="21b0a5f3-85fe-4f27-a139-3f1d21b063fa" containerName="pruner" Nov 23 06:48:38 crc kubenswrapper[4559]: E1123 06:48:38.251168 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25720f84-faae-4d25-8bac-7942852c0a28" containerName="extract-content" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.251172 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="25720f84-faae-4d25-8bac-7942852c0a28" containerName="extract-content" Nov 23 06:48:38 crc kubenswrapper[4559]: E1123 06:48:38.251179 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79f79383-e237-4748-bdc0-1edb62550872" containerName="extract-utilities" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.251185 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="79f79383-e237-4748-bdc0-1edb62550872" containerName="extract-utilities" Nov 23 06:48:38 crc kubenswrapper[4559]: E1123 06:48:38.251193 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25720f84-faae-4d25-8bac-7942852c0a28" containerName="extract-utilities" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.251200 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="25720f84-faae-4d25-8bac-7942852c0a28" containerName="extract-utilities" Nov 23 06:48:38 crc kubenswrapper[4559]: E1123 06:48:38.251208 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1d0d430-031f-448e-9763-69deb17abc4a" containerName="collect-profiles" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.251214 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1d0d430-031f-448e-9763-69deb17abc4a" containerName="collect-profiles" Nov 23 06:48:38 crc kubenswrapper[4559]: E1123 06:48:38.251219 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79f79383-e237-4748-bdc0-1edb62550872" containerName="registry-server" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.251225 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="79f79383-e237-4748-bdc0-1edb62550872" containerName="registry-server" Nov 23 06:48:38 crc kubenswrapper[4559]: E1123 06:48:38.251230 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7b5892a-8920-4cef-94a1-3736328e9a39" containerName="oauth-openshift" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.251235 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7b5892a-8920-4cef-94a1-3736328e9a39" containerName="oauth-openshift" Nov 23 06:48:38 crc kubenswrapper[4559]: E1123 06:48:38.251242 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79f79383-e237-4748-bdc0-1edb62550872" containerName="extract-content" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.251248 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="79f79383-e237-4748-bdc0-1edb62550872" containerName="extract-content" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.251337 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="eccd095a-81da-4c51-8107-0ad84257c648" containerName="pruner" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.251346 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7b5892a-8920-4cef-94a1-3736328e9a39" containerName="oauth-openshift" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.251352 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="7848f717-c7e3-40df-a4d8-854794e3c2e8" containerName="registry-server" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.251360 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="79f79383-e237-4748-bdc0-1edb62550872" containerName="registry-server" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.251368 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1d0d430-031f-448e-9763-69deb17abc4a" containerName="collect-profiles" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.251374 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="21b0a5f3-85fe-4f27-a139-3f1d21b063fa" containerName="pruner" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.251380 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="25720f84-faae-4d25-8bac-7942852c0a28" containerName="registry-server" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.251385 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="88092487-aea4-46ae-8918-35e1adf96708" containerName="registry-server" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.251715 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.255954 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e7b5892a-8920-4cef-94a1-3736328e9a39-audit-policies\") pod \"e7b5892a-8920-4cef-94a1-3736328e9a39\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.255996 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-service-ca\") pod \"e7b5892a-8920-4cef-94a1-3736328e9a39\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.256016 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-serving-cert\") pod \"e7b5892a-8920-4cef-94a1-3736328e9a39\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.256038 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-ocp-branding-template\") pod \"e7b5892a-8920-4cef-94a1-3736328e9a39\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.256069 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-router-certs\") pod \"e7b5892a-8920-4cef-94a1-3736328e9a39\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.256084 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-user-idp-0-file-data\") pod \"e7b5892a-8920-4cef-94a1-3736328e9a39\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.256106 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-user-template-provider-selection\") pod \"e7b5892a-8920-4cef-94a1-3736328e9a39\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.256131 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e7b5892a-8920-4cef-94a1-3736328e9a39-audit-dir\") pod \"e7b5892a-8920-4cef-94a1-3736328e9a39\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.256152 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-user-template-login\") pod \"e7b5892a-8920-4cef-94a1-3736328e9a39\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.256167 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-cliconfig\") pod \"e7b5892a-8920-4cef-94a1-3736328e9a39\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.256185 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-trusted-ca-bundle\") pod \"e7b5892a-8920-4cef-94a1-3736328e9a39\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.256205 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-session\") pod \"e7b5892a-8920-4cef-94a1-3736328e9a39\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.256245 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z57mp\" (UniqueName: \"kubernetes.io/projected/e7b5892a-8920-4cef-94a1-3736328e9a39-kube-api-access-z57mp\") pod \"e7b5892a-8920-4cef-94a1-3736328e9a39\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.256263 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-user-template-error\") pod \"e7b5892a-8920-4cef-94a1-3736328e9a39\" (UID: \"e7b5892a-8920-4cef-94a1-3736328e9a39\") " Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.256384 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.256412 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.256431 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ccb85a62-6a2d-452d-b625-28348eba96a1-audit-policies\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.256445 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-user-template-login\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.256462 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-user-template-error\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.256480 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.256515 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-system-service-ca\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.256553 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-system-router-certs\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.256579 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zh67x\" (UniqueName: \"kubernetes.io/projected/ccb85a62-6a2d-452d-b625-28348eba96a1-kube-api-access-zh67x\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.256611 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.256639 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.256683 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ccb85a62-6a2d-452d-b625-28348eba96a1-audit-dir\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.256704 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-system-session\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.256717 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.257360 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7b5892a-8920-4cef-94a1-3736328e9a39-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "e7b5892a-8920-4cef-94a1-3736328e9a39" (UID: "e7b5892a-8920-4cef-94a1-3736328e9a39"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.258606 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7c47d65989-g8qpb"] Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.258702 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "e7b5892a-8920-4cef-94a1-3736328e9a39" (UID: "e7b5892a-8920-4cef-94a1-3736328e9a39"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.259031 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e7b5892a-8920-4cef-94a1-3736328e9a39-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "e7b5892a-8920-4cef-94a1-3736328e9a39" (UID: "e7b5892a-8920-4cef-94a1-3736328e9a39"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.259139 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "e7b5892a-8920-4cef-94a1-3736328e9a39" (UID: "e7b5892a-8920-4cef-94a1-3736328e9a39"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.260955 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "e7b5892a-8920-4cef-94a1-3736328e9a39" (UID: "e7b5892a-8920-4cef-94a1-3736328e9a39"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.263871 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "e7b5892a-8920-4cef-94a1-3736328e9a39" (UID: "e7b5892a-8920-4cef-94a1-3736328e9a39"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.276764 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "e7b5892a-8920-4cef-94a1-3736328e9a39" (UID: "e7b5892a-8920-4cef-94a1-3736328e9a39"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.277124 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "e7b5892a-8920-4cef-94a1-3736328e9a39" (UID: "e7b5892a-8920-4cef-94a1-3736328e9a39"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.277325 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7b5892a-8920-4cef-94a1-3736328e9a39-kube-api-access-z57mp" (OuterVolumeSpecName: "kube-api-access-z57mp") pod "e7b5892a-8920-4cef-94a1-3736328e9a39" (UID: "e7b5892a-8920-4cef-94a1-3736328e9a39"). InnerVolumeSpecName "kube-api-access-z57mp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.277486 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "e7b5892a-8920-4cef-94a1-3736328e9a39" (UID: "e7b5892a-8920-4cef-94a1-3736328e9a39"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.277627 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "e7b5892a-8920-4cef-94a1-3736328e9a39" (UID: "e7b5892a-8920-4cef-94a1-3736328e9a39"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.277824 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "e7b5892a-8920-4cef-94a1-3736328e9a39" (UID: "e7b5892a-8920-4cef-94a1-3736328e9a39"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.277888 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "e7b5892a-8920-4cef-94a1-3736328e9a39" (UID: "e7b5892a-8920-4cef-94a1-3736328e9a39"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.278059 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "e7b5892a-8920-4cef-94a1-3736328e9a39" (UID: "e7b5892a-8920-4cef-94a1-3736328e9a39"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.357269 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-system-router-certs\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.357308 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zh67x\" (UniqueName: \"kubernetes.io/projected/ccb85a62-6a2d-452d-b625-28348eba96a1-kube-api-access-zh67x\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.357354 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.357411 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.357429 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ccb85a62-6a2d-452d-b625-28348eba96a1-audit-dir\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.357446 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-system-session\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.357461 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.357480 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.357506 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.357522 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-user-template-login\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.357544 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ccb85a62-6a2d-452d-b625-28348eba96a1-audit-policies\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.357575 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-user-template-error\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.357592 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.357629 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-system-service-ca\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.357686 4559 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.357696 4559 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.357705 4559 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.357715 4559 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.357731 4559 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.357741 4559 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.357751 4559 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e7b5892a-8920-4cef-94a1-3736328e9a39-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.357759 4559 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.357767 4559 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.357776 4559 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.357785 4559 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.357794 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z57mp\" (UniqueName: \"kubernetes.io/projected/e7b5892a-8920-4cef-94a1-3736328e9a39-kube-api-access-z57mp\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.357802 4559 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e7b5892a-8920-4cef-94a1-3736328e9a39-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.357810 4559 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e7b5892a-8920-4cef-94a1-3736328e9a39-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.358356 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-system-service-ca\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.358795 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ccb85a62-6a2d-452d-b625-28348eba96a1-audit-dir\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.358865 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.358955 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ccb85a62-6a2d-452d-b625-28348eba96a1-audit-policies\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.359265 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.360934 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-user-template-login\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.360972 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.361596 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-system-session\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.361696 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-system-router-certs\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.361791 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.361926 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-user-template-error\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.362232 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.362374 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ccb85a62-6a2d-452d-b625-28348eba96a1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.372381 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zh67x\" (UniqueName: \"kubernetes.io/projected/ccb85a62-6a2d-452d-b625-28348eba96a1-kube-api-access-zh67x\") pod \"oauth-openshift-7c47d65989-g8qpb\" (UID: \"ccb85a62-6a2d-452d-b625-28348eba96a1\") " pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.528548 4559 generic.go:334] "Generic (PLEG): container finished" podID="e7b5892a-8920-4cef-94a1-3736328e9a39" containerID="ce69456f4d4198e8b9edc01bd976a0de7270cf5baf44e52b38599f01a9ab6c82" exitCode=0 Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.528608 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.528622 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" event={"ID":"e7b5892a-8920-4cef-94a1-3736328e9a39","Type":"ContainerDied","Data":"ce69456f4d4198e8b9edc01bd976a0de7270cf5baf44e52b38599f01a9ab6c82"} Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.529127 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-kg77n" event={"ID":"e7b5892a-8920-4cef-94a1-3736328e9a39","Type":"ContainerDied","Data":"dae9c05d7f770dec3301fbffab75de3359ae8a0a030639f242ee6c49d68826b6"} Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.529199 4559 scope.go:117] "RemoveContainer" containerID="ce69456f4d4198e8b9edc01bd976a0de7270cf5baf44e52b38599f01a9ab6c82" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.539781 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-kg77n"] Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.541406 4559 scope.go:117] "RemoveContainer" containerID="ce69456f4d4198e8b9edc01bd976a0de7270cf5baf44e52b38599f01a9ab6c82" Nov 23 06:48:38 crc kubenswrapper[4559]: E1123 06:48:38.541705 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce69456f4d4198e8b9edc01bd976a0de7270cf5baf44e52b38599f01a9ab6c82\": container with ID starting with ce69456f4d4198e8b9edc01bd976a0de7270cf5baf44e52b38599f01a9ab6c82 not found: ID does not exist" containerID="ce69456f4d4198e8b9edc01bd976a0de7270cf5baf44e52b38599f01a9ab6c82" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.541797 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce69456f4d4198e8b9edc01bd976a0de7270cf5baf44e52b38599f01a9ab6c82"} err="failed to get container status \"ce69456f4d4198e8b9edc01bd976a0de7270cf5baf44e52b38599f01a9ab6c82\": rpc error: code = NotFound desc = could not find container \"ce69456f4d4198e8b9edc01bd976a0de7270cf5baf44e52b38599f01a9ab6c82\": container with ID starting with ce69456f4d4198e8b9edc01bd976a0de7270cf5baf44e52b38599f01a9ab6c82 not found: ID does not exist" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.542305 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-kg77n"] Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.588091 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:38 crc kubenswrapper[4559]: I1123 06:48:38.912955 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7c47d65989-g8qpb"] Nov 23 06:48:39 crc kubenswrapper[4559]: I1123 06:48:39.535022 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" event={"ID":"ccb85a62-6a2d-452d-b625-28348eba96a1","Type":"ContainerStarted","Data":"2b05b4b2163ca6abc54ba400f7e8f1120644eb1a2935a8a2886dbf711826c041"} Nov 23 06:48:39 crc kubenswrapper[4559]: I1123 06:48:39.535185 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" event={"ID":"ccb85a62-6a2d-452d-b625-28348eba96a1","Type":"ContainerStarted","Data":"abbdccd104fea0fa8111c8d520683fcfcef31cf91afe3f9660f78a6c63820b0b"} Nov 23 06:48:39 crc kubenswrapper[4559]: I1123 06:48:39.535399 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:39 crc kubenswrapper[4559]: I1123 06:48:39.538867 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" Nov 23 06:48:39 crc kubenswrapper[4559]: I1123 06:48:39.563202 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7c47d65989-g8qpb" podStartSLOduration=27.56318578 podStartE2EDuration="27.56318578s" podCreationTimestamp="2025-11-23 06:48:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:48:39.548056722 +0000 UTC m=+221.570042337" watchObservedRunningTime="2025-11-23 06:48:39.56318578 +0000 UTC m=+221.585171383" Nov 23 06:48:40 crc kubenswrapper[4559]: I1123 06:48:40.281863 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7b5892a-8920-4cef-94a1-3736328e9a39" path="/var/lib/kubelet/pods/e7b5892a-8920-4cef-94a1-3736328e9a39/volumes" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.186811 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qhvl6"] Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.187368 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qhvl6" podUID="680dc567-9fb8-4396-adb4-76bb44250c9c" containerName="registry-server" containerID="cri-o://bf6798357df248f446eddbc97e4fec2a6e2797156dd221864a7f1ef145ef23df" gracePeriod=30 Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.195790 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-whrv6"] Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.198203 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-whrv6" podUID="9d1c86c0-8a6f-4914-9593-20ca1220ed8b" containerName="registry-server" containerID="cri-o://583d1a00ef704e1ed48d88680d0fd97403d80d0a5354bf15dbb8f6eaba7d20e0" gracePeriod=30 Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.198674 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tgbrt"] Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.198827 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-tgbrt" podUID="337f295d-6c74-4850-bb14-125192de4385" containerName="marketplace-operator" containerID="cri-o://8381279476e92624c58b3ea09f34af46e6f5c8c4061b9534c63f0d9d299c6df3" gracePeriod=30 Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.207434 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m9ltm"] Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.207822 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-m9ltm" podUID="98640d4a-0cbc-428c-8f7a-25a1b36f16f9" containerName="registry-server" containerID="cri-o://6518fe7506c3af453b8f7ee9cbbe1235ad62a2c03155cb4a853bfc176e155b33" gracePeriod=30 Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.216695 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v7hpt"] Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.217286 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-v7hpt" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.218172 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rz5lj"] Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.218349 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rz5lj" podUID="f59cbebb-100b-4d79-83f8-bf1a625caf93" containerName="registry-server" containerID="cri-o://73b67ebd55b02d6deacf6718d5f6047c50a55e888094a012afe1ccac9b3fdd9e" gracePeriod=30 Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.225473 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v7hpt"] Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.226071 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/351f8b04-cba3-4484-94b6-c4abd1c9be42-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-v7hpt\" (UID: \"351f8b04-cba3-4484-94b6-c4abd1c9be42\") " pod="openshift-marketplace/marketplace-operator-79b997595-v7hpt" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.226104 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/351f8b04-cba3-4484-94b6-c4abd1c9be42-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-v7hpt\" (UID: \"351f8b04-cba3-4484-94b6-c4abd1c9be42\") " pod="openshift-marketplace/marketplace-operator-79b997595-v7hpt" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.226122 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9ph2\" (UniqueName: \"kubernetes.io/projected/351f8b04-cba3-4484-94b6-c4abd1c9be42-kube-api-access-g9ph2\") pod \"marketplace-operator-79b997595-v7hpt\" (UID: \"351f8b04-cba3-4484-94b6-c4abd1c9be42\") " pod="openshift-marketplace/marketplace-operator-79b997595-v7hpt" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.327005 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/351f8b04-cba3-4484-94b6-c4abd1c9be42-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-v7hpt\" (UID: \"351f8b04-cba3-4484-94b6-c4abd1c9be42\") " pod="openshift-marketplace/marketplace-operator-79b997595-v7hpt" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.327216 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/351f8b04-cba3-4484-94b6-c4abd1c9be42-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-v7hpt\" (UID: \"351f8b04-cba3-4484-94b6-c4abd1c9be42\") " pod="openshift-marketplace/marketplace-operator-79b997595-v7hpt" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.327236 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9ph2\" (UniqueName: \"kubernetes.io/projected/351f8b04-cba3-4484-94b6-c4abd1c9be42-kube-api-access-g9ph2\") pod \"marketplace-operator-79b997595-v7hpt\" (UID: \"351f8b04-cba3-4484-94b6-c4abd1c9be42\") " pod="openshift-marketplace/marketplace-operator-79b997595-v7hpt" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.328172 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/351f8b04-cba3-4484-94b6-c4abd1c9be42-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-v7hpt\" (UID: \"351f8b04-cba3-4484-94b6-c4abd1c9be42\") " pod="openshift-marketplace/marketplace-operator-79b997595-v7hpt" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.331900 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/351f8b04-cba3-4484-94b6-c4abd1c9be42-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-v7hpt\" (UID: \"351f8b04-cba3-4484-94b6-c4abd1c9be42\") " pod="openshift-marketplace/marketplace-operator-79b997595-v7hpt" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.340195 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9ph2\" (UniqueName: \"kubernetes.io/projected/351f8b04-cba3-4484-94b6-c4abd1c9be42-kube-api-access-g9ph2\") pod \"marketplace-operator-79b997595-v7hpt\" (UID: \"351f8b04-cba3-4484-94b6-c4abd1c9be42\") " pod="openshift-marketplace/marketplace-operator-79b997595-v7hpt" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.598170 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-v7hpt" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.601852 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qhvl6" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.606342 4559 generic.go:334] "Generic (PLEG): container finished" podID="9d1c86c0-8a6f-4914-9593-20ca1220ed8b" containerID="583d1a00ef704e1ed48d88680d0fd97403d80d0a5354bf15dbb8f6eaba7d20e0" exitCode=0 Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.606405 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whrv6" event={"ID":"9d1c86c0-8a6f-4914-9593-20ca1220ed8b","Type":"ContainerDied","Data":"583d1a00ef704e1ed48d88680d0fd97403d80d0a5354bf15dbb8f6eaba7d20e0"} Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.606415 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tgbrt" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.606431 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whrv6" event={"ID":"9d1c86c0-8a6f-4914-9593-20ca1220ed8b","Type":"ContainerDied","Data":"e01d714e3a17603b995e082d90b605afc5bd1651356adda9586662a3517766c7"} Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.606442 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e01d714e3a17603b995e082d90b605afc5bd1651356adda9586662a3517766c7" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.608556 4559 generic.go:334] "Generic (PLEG): container finished" podID="337f295d-6c74-4850-bb14-125192de4385" containerID="8381279476e92624c58b3ea09f34af46e6f5c8c4061b9534c63f0d9d299c6df3" exitCode=0 Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.608602 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tgbrt" event={"ID":"337f295d-6c74-4850-bb14-125192de4385","Type":"ContainerDied","Data":"8381279476e92624c58b3ea09f34af46e6f5c8c4061b9534c63f0d9d299c6df3"} Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.608619 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tgbrt" event={"ID":"337f295d-6c74-4850-bb14-125192de4385","Type":"ContainerDied","Data":"43e164f17fcf543c727f127b9a2bab378a33f398a4a95311b501ec89ed31ad3b"} Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.608634 4559 scope.go:117] "RemoveContainer" containerID="8381279476e92624c58b3ea09f34af46e6f5c8c4061b9534c63f0d9d299c6df3" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.612033 4559 generic.go:334] "Generic (PLEG): container finished" podID="680dc567-9fb8-4396-adb4-76bb44250c9c" containerID="bf6798357df248f446eddbc97e4fec2a6e2797156dd221864a7f1ef145ef23df" exitCode=0 Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.612070 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qhvl6" event={"ID":"680dc567-9fb8-4396-adb4-76bb44250c9c","Type":"ContainerDied","Data":"bf6798357df248f446eddbc97e4fec2a6e2797156dd221864a7f1ef145ef23df"} Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.612102 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qhvl6" event={"ID":"680dc567-9fb8-4396-adb4-76bb44250c9c","Type":"ContainerDied","Data":"8a25804bc3e1fcd40c2a4a9d074af5b808167846082d5116deca070c9c793622"} Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.612115 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qhvl6" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.615893 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m9ltm" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.618427 4559 generic.go:334] "Generic (PLEG): container finished" podID="f59cbebb-100b-4d79-83f8-bf1a625caf93" containerID="73b67ebd55b02d6deacf6718d5f6047c50a55e888094a012afe1ccac9b3fdd9e" exitCode=0 Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.618476 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rz5lj" event={"ID":"f59cbebb-100b-4d79-83f8-bf1a625caf93","Type":"ContainerDied","Data":"73b67ebd55b02d6deacf6718d5f6047c50a55e888094a012afe1ccac9b3fdd9e"} Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.618501 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rz5lj" event={"ID":"f59cbebb-100b-4d79-83f8-bf1a625caf93","Type":"ContainerDied","Data":"8f11e6da0278ec4a8b7586a1d8d6df08e9f796ba7b0314b6043d34a838b88db9"} Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.618512 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f11e6da0278ec4a8b7586a1d8d6df08e9f796ba7b0314b6043d34a838b88db9" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.623249 4559 generic.go:334] "Generic (PLEG): container finished" podID="98640d4a-0cbc-428c-8f7a-25a1b36f16f9" containerID="6518fe7506c3af453b8f7ee9cbbe1235ad62a2c03155cb4a853bfc176e155b33" exitCode=0 Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.623281 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m9ltm" event={"ID":"98640d4a-0cbc-428c-8f7a-25a1b36f16f9","Type":"ContainerDied","Data":"6518fe7506c3af453b8f7ee9cbbe1235ad62a2c03155cb4a853bfc176e155b33"} Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.623290 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m9ltm" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.623300 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m9ltm" event={"ID":"98640d4a-0cbc-428c-8f7a-25a1b36f16f9","Type":"ContainerDied","Data":"c48ff1f993a90e1623d868ab760bb058ed436429c40f6a9910351988e590c75b"} Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.624355 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-whrv6" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.628662 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rz5lj" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.634529 4559 scope.go:117] "RemoveContainer" containerID="8381279476e92624c58b3ea09f34af46e6f5c8c4061b9534c63f0d9d299c6df3" Nov 23 06:48:56 crc kubenswrapper[4559]: E1123 06:48:56.634795 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8381279476e92624c58b3ea09f34af46e6f5c8c4061b9534c63f0d9d299c6df3\": container with ID starting with 8381279476e92624c58b3ea09f34af46e6f5c8c4061b9534c63f0d9d299c6df3 not found: ID does not exist" containerID="8381279476e92624c58b3ea09f34af46e6f5c8c4061b9534c63f0d9d299c6df3" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.634827 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8381279476e92624c58b3ea09f34af46e6f5c8c4061b9534c63f0d9d299c6df3"} err="failed to get container status \"8381279476e92624c58b3ea09f34af46e6f5c8c4061b9534c63f0d9d299c6df3\": rpc error: code = NotFound desc = could not find container \"8381279476e92624c58b3ea09f34af46e6f5c8c4061b9534c63f0d9d299c6df3\": container with ID starting with 8381279476e92624c58b3ea09f34af46e6f5c8c4061b9534c63f0d9d299c6df3 not found: ID does not exist" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.634849 4559 scope.go:117] "RemoveContainer" containerID="bf6798357df248f446eddbc97e4fec2a6e2797156dd221864a7f1ef145ef23df" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.653013 4559 scope.go:117] "RemoveContainer" containerID="b06e34d44c1451202da39a5592c3d490ef6b02a8f069a17d1fc1f4329c38022a" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.681993 4559 scope.go:117] "RemoveContainer" containerID="270ed989791553c1822cfb134bdec27290409b476d30e7e2c5e95ed1da292bb9" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.708850 4559 scope.go:117] "RemoveContainer" containerID="bf6798357df248f446eddbc97e4fec2a6e2797156dd221864a7f1ef145ef23df" Nov 23 06:48:56 crc kubenswrapper[4559]: E1123 06:48:56.709528 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf6798357df248f446eddbc97e4fec2a6e2797156dd221864a7f1ef145ef23df\": container with ID starting with bf6798357df248f446eddbc97e4fec2a6e2797156dd221864a7f1ef145ef23df not found: ID does not exist" containerID="bf6798357df248f446eddbc97e4fec2a6e2797156dd221864a7f1ef145ef23df" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.709554 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf6798357df248f446eddbc97e4fec2a6e2797156dd221864a7f1ef145ef23df"} err="failed to get container status \"bf6798357df248f446eddbc97e4fec2a6e2797156dd221864a7f1ef145ef23df\": rpc error: code = NotFound desc = could not find container \"bf6798357df248f446eddbc97e4fec2a6e2797156dd221864a7f1ef145ef23df\": container with ID starting with bf6798357df248f446eddbc97e4fec2a6e2797156dd221864a7f1ef145ef23df not found: ID does not exist" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.709578 4559 scope.go:117] "RemoveContainer" containerID="b06e34d44c1451202da39a5592c3d490ef6b02a8f069a17d1fc1f4329c38022a" Nov 23 06:48:56 crc kubenswrapper[4559]: E1123 06:48:56.719266 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b06e34d44c1451202da39a5592c3d490ef6b02a8f069a17d1fc1f4329c38022a\": container with ID starting with b06e34d44c1451202da39a5592c3d490ef6b02a8f069a17d1fc1f4329c38022a not found: ID does not exist" containerID="b06e34d44c1451202da39a5592c3d490ef6b02a8f069a17d1fc1f4329c38022a" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.719297 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b06e34d44c1451202da39a5592c3d490ef6b02a8f069a17d1fc1f4329c38022a"} err="failed to get container status \"b06e34d44c1451202da39a5592c3d490ef6b02a8f069a17d1fc1f4329c38022a\": rpc error: code = NotFound desc = could not find container \"b06e34d44c1451202da39a5592c3d490ef6b02a8f069a17d1fc1f4329c38022a\": container with ID starting with b06e34d44c1451202da39a5592c3d490ef6b02a8f069a17d1fc1f4329c38022a not found: ID does not exist" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.719318 4559 scope.go:117] "RemoveContainer" containerID="270ed989791553c1822cfb134bdec27290409b476d30e7e2c5e95ed1da292bb9" Nov 23 06:48:56 crc kubenswrapper[4559]: E1123 06:48:56.719909 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"270ed989791553c1822cfb134bdec27290409b476d30e7e2c5e95ed1da292bb9\": container with ID starting with 270ed989791553c1822cfb134bdec27290409b476d30e7e2c5e95ed1da292bb9 not found: ID does not exist" containerID="270ed989791553c1822cfb134bdec27290409b476d30e7e2c5e95ed1da292bb9" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.719950 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"270ed989791553c1822cfb134bdec27290409b476d30e7e2c5e95ed1da292bb9"} err="failed to get container status \"270ed989791553c1822cfb134bdec27290409b476d30e7e2c5e95ed1da292bb9\": rpc error: code = NotFound desc = could not find container \"270ed989791553c1822cfb134bdec27290409b476d30e7e2c5e95ed1da292bb9\": container with ID starting with 270ed989791553c1822cfb134bdec27290409b476d30e7e2c5e95ed1da292bb9 not found: ID does not exist" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.719970 4559 scope.go:117] "RemoveContainer" containerID="6518fe7506c3af453b8f7ee9cbbe1235ad62a2c03155cb4a853bfc176e155b33" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.732524 4559 scope.go:117] "RemoveContainer" containerID="ab8cc8db23b649332b5cfa20be82533ab922e3841569a311ea5d8f5111ea1296" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.732782 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d1c86c0-8a6f-4914-9593-20ca1220ed8b-utilities\") pod \"9d1c86c0-8a6f-4914-9593-20ca1220ed8b\" (UID: \"9d1c86c0-8a6f-4914-9593-20ca1220ed8b\") " Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.732831 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dktv\" (UniqueName: \"kubernetes.io/projected/680dc567-9fb8-4396-adb4-76bb44250c9c-kube-api-access-7dktv\") pod \"680dc567-9fb8-4396-adb4-76bb44250c9c\" (UID: \"680dc567-9fb8-4396-adb4-76bb44250c9c\") " Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.732857 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/337f295d-6c74-4850-bb14-125192de4385-marketplace-operator-metrics\") pod \"337f295d-6c74-4850-bb14-125192de4385\" (UID: \"337f295d-6c74-4850-bb14-125192de4385\") " Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.732874 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gh72\" (UniqueName: \"kubernetes.io/projected/9d1c86c0-8a6f-4914-9593-20ca1220ed8b-kube-api-access-9gh72\") pod \"9d1c86c0-8a6f-4914-9593-20ca1220ed8b\" (UID: \"9d1c86c0-8a6f-4914-9593-20ca1220ed8b\") " Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.732916 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxpff\" (UniqueName: \"kubernetes.io/projected/98640d4a-0cbc-428c-8f7a-25a1b36f16f9-kube-api-access-lxpff\") pod \"98640d4a-0cbc-428c-8f7a-25a1b36f16f9\" (UID: \"98640d4a-0cbc-428c-8f7a-25a1b36f16f9\") " Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.732936 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d1c86c0-8a6f-4914-9593-20ca1220ed8b-catalog-content\") pod \"9d1c86c0-8a6f-4914-9593-20ca1220ed8b\" (UID: \"9d1c86c0-8a6f-4914-9593-20ca1220ed8b\") " Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.732972 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/680dc567-9fb8-4396-adb4-76bb44250c9c-utilities\") pod \"680dc567-9fb8-4396-adb4-76bb44250c9c\" (UID: \"680dc567-9fb8-4396-adb4-76bb44250c9c\") " Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.733020 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98640d4a-0cbc-428c-8f7a-25a1b36f16f9-catalog-content\") pod \"98640d4a-0cbc-428c-8f7a-25a1b36f16f9\" (UID: \"98640d4a-0cbc-428c-8f7a-25a1b36f16f9\") " Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.733062 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f59cbebb-100b-4d79-83f8-bf1a625caf93-utilities\") pod \"f59cbebb-100b-4d79-83f8-bf1a625caf93\" (UID: \"f59cbebb-100b-4d79-83f8-bf1a625caf93\") " Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.733080 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/337f295d-6c74-4850-bb14-125192de4385-marketplace-trusted-ca\") pod \"337f295d-6c74-4850-bb14-125192de4385\" (UID: \"337f295d-6c74-4850-bb14-125192de4385\") " Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.733879 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d1c86c0-8a6f-4914-9593-20ca1220ed8b-utilities" (OuterVolumeSpecName: "utilities") pod "9d1c86c0-8a6f-4914-9593-20ca1220ed8b" (UID: "9d1c86c0-8a6f-4914-9593-20ca1220ed8b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.734068 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/680dc567-9fb8-4396-adb4-76bb44250c9c-utilities" (OuterVolumeSpecName: "utilities") pod "680dc567-9fb8-4396-adb4-76bb44250c9c" (UID: "680dc567-9fb8-4396-adb4-76bb44250c9c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.734292 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f59cbebb-100b-4d79-83f8-bf1a625caf93-utilities" (OuterVolumeSpecName: "utilities") pod "f59cbebb-100b-4d79-83f8-bf1a625caf93" (UID: "f59cbebb-100b-4d79-83f8-bf1a625caf93"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.734502 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqq59\" (UniqueName: \"kubernetes.io/projected/337f295d-6c74-4850-bb14-125192de4385-kube-api-access-pqq59\") pod \"337f295d-6c74-4850-bb14-125192de4385\" (UID: \"337f295d-6c74-4850-bb14-125192de4385\") " Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.734543 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98640d4a-0cbc-428c-8f7a-25a1b36f16f9-utilities\") pod \"98640d4a-0cbc-428c-8f7a-25a1b36f16f9\" (UID: \"98640d4a-0cbc-428c-8f7a-25a1b36f16f9\") " Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.734595 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmnvz\" (UniqueName: \"kubernetes.io/projected/f59cbebb-100b-4d79-83f8-bf1a625caf93-kube-api-access-vmnvz\") pod \"f59cbebb-100b-4d79-83f8-bf1a625caf93\" (UID: \"f59cbebb-100b-4d79-83f8-bf1a625caf93\") " Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.734617 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/680dc567-9fb8-4396-adb4-76bb44250c9c-catalog-content\") pod \"680dc567-9fb8-4396-adb4-76bb44250c9c\" (UID: \"680dc567-9fb8-4396-adb4-76bb44250c9c\") " Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.734636 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f59cbebb-100b-4d79-83f8-bf1a625caf93-catalog-content\") pod \"f59cbebb-100b-4d79-83f8-bf1a625caf93\" (UID: \"f59cbebb-100b-4d79-83f8-bf1a625caf93\") " Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.734964 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d1c86c0-8a6f-4914-9593-20ca1220ed8b-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.734983 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/680dc567-9fb8-4396-adb4-76bb44250c9c-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.734991 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f59cbebb-100b-4d79-83f8-bf1a625caf93-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.735835 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98640d4a-0cbc-428c-8f7a-25a1b36f16f9-utilities" (OuterVolumeSpecName: "utilities") pod "98640d4a-0cbc-428c-8f7a-25a1b36f16f9" (UID: "98640d4a-0cbc-428c-8f7a-25a1b36f16f9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.736038 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/337f295d-6c74-4850-bb14-125192de4385-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "337f295d-6c74-4850-bb14-125192de4385" (UID: "337f295d-6c74-4850-bb14-125192de4385"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.738285 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/680dc567-9fb8-4396-adb4-76bb44250c9c-kube-api-access-7dktv" (OuterVolumeSpecName: "kube-api-access-7dktv") pod "680dc567-9fb8-4396-adb4-76bb44250c9c" (UID: "680dc567-9fb8-4396-adb4-76bb44250c9c"). InnerVolumeSpecName "kube-api-access-7dktv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.738918 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d1c86c0-8a6f-4914-9593-20ca1220ed8b-kube-api-access-9gh72" (OuterVolumeSpecName: "kube-api-access-9gh72") pod "9d1c86c0-8a6f-4914-9593-20ca1220ed8b" (UID: "9d1c86c0-8a6f-4914-9593-20ca1220ed8b"). InnerVolumeSpecName "kube-api-access-9gh72". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.739302 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/337f295d-6c74-4850-bb14-125192de4385-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "337f295d-6c74-4850-bb14-125192de4385" (UID: "337f295d-6c74-4850-bb14-125192de4385"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.739903 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f59cbebb-100b-4d79-83f8-bf1a625caf93-kube-api-access-vmnvz" (OuterVolumeSpecName: "kube-api-access-vmnvz") pod "f59cbebb-100b-4d79-83f8-bf1a625caf93" (UID: "f59cbebb-100b-4d79-83f8-bf1a625caf93"). InnerVolumeSpecName "kube-api-access-vmnvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.739976 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/337f295d-6c74-4850-bb14-125192de4385-kube-api-access-pqq59" (OuterVolumeSpecName: "kube-api-access-pqq59") pod "337f295d-6c74-4850-bb14-125192de4385" (UID: "337f295d-6c74-4850-bb14-125192de4385"). InnerVolumeSpecName "kube-api-access-pqq59". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.742201 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98640d4a-0cbc-428c-8f7a-25a1b36f16f9-kube-api-access-lxpff" (OuterVolumeSpecName: "kube-api-access-lxpff") pod "98640d4a-0cbc-428c-8f7a-25a1b36f16f9" (UID: "98640d4a-0cbc-428c-8f7a-25a1b36f16f9"). InnerVolumeSpecName "kube-api-access-lxpff". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.749649 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98640d4a-0cbc-428c-8f7a-25a1b36f16f9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "98640d4a-0cbc-428c-8f7a-25a1b36f16f9" (UID: "98640d4a-0cbc-428c-8f7a-25a1b36f16f9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.760092 4559 scope.go:117] "RemoveContainer" containerID="a7938a6b9af471f28df9be4c1aad4fedc267f6c68e3d8df79902a4c8d7100fe0" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.776038 4559 scope.go:117] "RemoveContainer" containerID="6518fe7506c3af453b8f7ee9cbbe1235ad62a2c03155cb4a853bfc176e155b33" Nov 23 06:48:56 crc kubenswrapper[4559]: E1123 06:48:56.776564 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6518fe7506c3af453b8f7ee9cbbe1235ad62a2c03155cb4a853bfc176e155b33\": container with ID starting with 6518fe7506c3af453b8f7ee9cbbe1235ad62a2c03155cb4a853bfc176e155b33 not found: ID does not exist" containerID="6518fe7506c3af453b8f7ee9cbbe1235ad62a2c03155cb4a853bfc176e155b33" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.776601 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6518fe7506c3af453b8f7ee9cbbe1235ad62a2c03155cb4a853bfc176e155b33"} err="failed to get container status \"6518fe7506c3af453b8f7ee9cbbe1235ad62a2c03155cb4a853bfc176e155b33\": rpc error: code = NotFound desc = could not find container \"6518fe7506c3af453b8f7ee9cbbe1235ad62a2c03155cb4a853bfc176e155b33\": container with ID starting with 6518fe7506c3af453b8f7ee9cbbe1235ad62a2c03155cb4a853bfc176e155b33 not found: ID does not exist" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.776627 4559 scope.go:117] "RemoveContainer" containerID="ab8cc8db23b649332b5cfa20be82533ab922e3841569a311ea5d8f5111ea1296" Nov 23 06:48:56 crc kubenswrapper[4559]: E1123 06:48:56.777042 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab8cc8db23b649332b5cfa20be82533ab922e3841569a311ea5d8f5111ea1296\": container with ID starting with ab8cc8db23b649332b5cfa20be82533ab922e3841569a311ea5d8f5111ea1296 not found: ID does not exist" containerID="ab8cc8db23b649332b5cfa20be82533ab922e3841569a311ea5d8f5111ea1296" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.777089 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab8cc8db23b649332b5cfa20be82533ab922e3841569a311ea5d8f5111ea1296"} err="failed to get container status \"ab8cc8db23b649332b5cfa20be82533ab922e3841569a311ea5d8f5111ea1296\": rpc error: code = NotFound desc = could not find container \"ab8cc8db23b649332b5cfa20be82533ab922e3841569a311ea5d8f5111ea1296\": container with ID starting with ab8cc8db23b649332b5cfa20be82533ab922e3841569a311ea5d8f5111ea1296 not found: ID does not exist" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.777144 4559 scope.go:117] "RemoveContainer" containerID="a7938a6b9af471f28df9be4c1aad4fedc267f6c68e3d8df79902a4c8d7100fe0" Nov 23 06:48:56 crc kubenswrapper[4559]: E1123 06:48:56.777706 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7938a6b9af471f28df9be4c1aad4fedc267f6c68e3d8df79902a4c8d7100fe0\": container with ID starting with a7938a6b9af471f28df9be4c1aad4fedc267f6c68e3d8df79902a4c8d7100fe0 not found: ID does not exist" containerID="a7938a6b9af471f28df9be4c1aad4fedc267f6c68e3d8df79902a4c8d7100fe0" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.777779 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7938a6b9af471f28df9be4c1aad4fedc267f6c68e3d8df79902a4c8d7100fe0"} err="failed to get container status \"a7938a6b9af471f28df9be4c1aad4fedc267f6c68e3d8df79902a4c8d7100fe0\": rpc error: code = NotFound desc = could not find container \"a7938a6b9af471f28df9be4c1aad4fedc267f6c68e3d8df79902a4c8d7100fe0\": container with ID starting with a7938a6b9af471f28df9be4c1aad4fedc267f6c68e3d8df79902a4c8d7100fe0 not found: ID does not exist" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.786232 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d1c86c0-8a6f-4914-9593-20ca1220ed8b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9d1c86c0-8a6f-4914-9593-20ca1220ed8b" (UID: "9d1c86c0-8a6f-4914-9593-20ca1220ed8b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.799325 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/680dc567-9fb8-4396-adb4-76bb44250c9c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "680dc567-9fb8-4396-adb4-76bb44250c9c" (UID: "680dc567-9fb8-4396-adb4-76bb44250c9c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.808548 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f59cbebb-100b-4d79-83f8-bf1a625caf93-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f59cbebb-100b-4d79-83f8-bf1a625caf93" (UID: "f59cbebb-100b-4d79-83f8-bf1a625caf93"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.836596 4559 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/337f295d-6c74-4850-bb14-125192de4385-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.836638 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqq59\" (UniqueName: \"kubernetes.io/projected/337f295d-6c74-4850-bb14-125192de4385-kube-api-access-pqq59\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.836662 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98640d4a-0cbc-428c-8f7a-25a1b36f16f9-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.836674 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmnvz\" (UniqueName: \"kubernetes.io/projected/f59cbebb-100b-4d79-83f8-bf1a625caf93-kube-api-access-vmnvz\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.836682 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/680dc567-9fb8-4396-adb4-76bb44250c9c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.836698 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f59cbebb-100b-4d79-83f8-bf1a625caf93-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.836705 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dktv\" (UniqueName: \"kubernetes.io/projected/680dc567-9fb8-4396-adb4-76bb44250c9c-kube-api-access-7dktv\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.836713 4559 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/337f295d-6c74-4850-bb14-125192de4385-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.836721 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gh72\" (UniqueName: \"kubernetes.io/projected/9d1c86c0-8a6f-4914-9593-20ca1220ed8b-kube-api-access-9gh72\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.836729 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxpff\" (UniqueName: \"kubernetes.io/projected/98640d4a-0cbc-428c-8f7a-25a1b36f16f9-kube-api-access-lxpff\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.836736 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d1c86c0-8a6f-4914-9593-20ca1220ed8b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.836743 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98640d4a-0cbc-428c-8f7a-25a1b36f16f9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.937130 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qhvl6"] Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.940072 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qhvl6"] Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.950579 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m9ltm"] Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.955460 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-m9ltm"] Nov 23 06:48:56 crc kubenswrapper[4559]: I1123 06:48:56.961262 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v7hpt"] Nov 23 06:48:57 crc kubenswrapper[4559]: I1123 06:48:57.629964 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-v7hpt" event={"ID":"351f8b04-cba3-4484-94b6-c4abd1c9be42","Type":"ContainerStarted","Data":"f79b3043885f4ed652b9f2840c8f1e3a183e905c672f76a9a575923b398fc630"} Nov 23 06:48:57 crc kubenswrapper[4559]: I1123 06:48:57.630022 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-v7hpt" Nov 23 06:48:57 crc kubenswrapper[4559]: I1123 06:48:57.630039 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-v7hpt" event={"ID":"351f8b04-cba3-4484-94b6-c4abd1c9be42","Type":"ContainerStarted","Data":"b63bf3794dc2c514a4b48c29856238f68ccbd8db5e1c671cab7fbe35e9fabf8f"} Nov 23 06:48:57 crc kubenswrapper[4559]: I1123 06:48:57.630966 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tgbrt" Nov 23 06:48:57 crc kubenswrapper[4559]: I1123 06:48:57.632212 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-whrv6" Nov 23 06:48:57 crc kubenswrapper[4559]: I1123 06:48:57.632840 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rz5lj" Nov 23 06:48:57 crc kubenswrapper[4559]: I1123 06:48:57.633098 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-v7hpt" Nov 23 06:48:57 crc kubenswrapper[4559]: I1123 06:48:57.642718 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-v7hpt" podStartSLOduration=1.6427016490000002 podStartE2EDuration="1.642701649s" podCreationTimestamp="2025-11-23 06:48:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:48:57.641671544 +0000 UTC m=+239.663657158" watchObservedRunningTime="2025-11-23 06:48:57.642701649 +0000 UTC m=+239.664687263" Nov 23 06:48:57 crc kubenswrapper[4559]: I1123 06:48:57.670221 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tgbrt"] Nov 23 06:48:57 crc kubenswrapper[4559]: I1123 06:48:57.678840 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tgbrt"] Nov 23 06:48:57 crc kubenswrapper[4559]: I1123 06:48:57.682214 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-whrv6"] Nov 23 06:48:57 crc kubenswrapper[4559]: I1123 06:48:57.685171 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-whrv6"] Nov 23 06:48:57 crc kubenswrapper[4559]: I1123 06:48:57.688129 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rz5lj"] Nov 23 06:48:57 crc kubenswrapper[4559]: I1123 06:48:57.690232 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rz5lj"] Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.196691 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4zrs4"] Nov 23 06:48:58 crc kubenswrapper[4559]: E1123 06:48:58.196893 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="680dc567-9fb8-4396-adb4-76bb44250c9c" containerName="extract-utilities" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.196911 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="680dc567-9fb8-4396-adb4-76bb44250c9c" containerName="extract-utilities" Nov 23 06:48:58 crc kubenswrapper[4559]: E1123 06:48:58.196920 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98640d4a-0cbc-428c-8f7a-25a1b36f16f9" containerName="extract-utilities" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.196926 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="98640d4a-0cbc-428c-8f7a-25a1b36f16f9" containerName="extract-utilities" Nov 23 06:48:58 crc kubenswrapper[4559]: E1123 06:48:58.196935 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d1c86c0-8a6f-4914-9593-20ca1220ed8b" containerName="extract-content" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.196940 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d1c86c0-8a6f-4914-9593-20ca1220ed8b" containerName="extract-content" Nov 23 06:48:58 crc kubenswrapper[4559]: E1123 06:48:58.196947 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f59cbebb-100b-4d79-83f8-bf1a625caf93" containerName="extract-content" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.196952 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="f59cbebb-100b-4d79-83f8-bf1a625caf93" containerName="extract-content" Nov 23 06:48:58 crc kubenswrapper[4559]: E1123 06:48:58.196959 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d1c86c0-8a6f-4914-9593-20ca1220ed8b" containerName="extract-utilities" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.196964 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d1c86c0-8a6f-4914-9593-20ca1220ed8b" containerName="extract-utilities" Nov 23 06:48:58 crc kubenswrapper[4559]: E1123 06:48:58.196971 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="680dc567-9fb8-4396-adb4-76bb44250c9c" containerName="registry-server" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.196977 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="680dc567-9fb8-4396-adb4-76bb44250c9c" containerName="registry-server" Nov 23 06:48:58 crc kubenswrapper[4559]: E1123 06:48:58.196987 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="680dc567-9fb8-4396-adb4-76bb44250c9c" containerName="extract-content" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.196992 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="680dc567-9fb8-4396-adb4-76bb44250c9c" containerName="extract-content" Nov 23 06:48:58 crc kubenswrapper[4559]: E1123 06:48:58.196998 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98640d4a-0cbc-428c-8f7a-25a1b36f16f9" containerName="registry-server" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.197004 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="98640d4a-0cbc-428c-8f7a-25a1b36f16f9" containerName="registry-server" Nov 23 06:48:58 crc kubenswrapper[4559]: E1123 06:48:58.197012 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d1c86c0-8a6f-4914-9593-20ca1220ed8b" containerName="registry-server" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.197018 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d1c86c0-8a6f-4914-9593-20ca1220ed8b" containerName="registry-server" Nov 23 06:48:58 crc kubenswrapper[4559]: E1123 06:48:58.197024 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="337f295d-6c74-4850-bb14-125192de4385" containerName="marketplace-operator" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.197030 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="337f295d-6c74-4850-bb14-125192de4385" containerName="marketplace-operator" Nov 23 06:48:58 crc kubenswrapper[4559]: E1123 06:48:58.197187 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f59cbebb-100b-4d79-83f8-bf1a625caf93" containerName="registry-server" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.197196 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="f59cbebb-100b-4d79-83f8-bf1a625caf93" containerName="registry-server" Nov 23 06:48:58 crc kubenswrapper[4559]: E1123 06:48:58.197207 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f59cbebb-100b-4d79-83f8-bf1a625caf93" containerName="extract-utilities" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.197213 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="f59cbebb-100b-4d79-83f8-bf1a625caf93" containerName="extract-utilities" Nov 23 06:48:58 crc kubenswrapper[4559]: E1123 06:48:58.197220 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98640d4a-0cbc-428c-8f7a-25a1b36f16f9" containerName="extract-content" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.197225 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="98640d4a-0cbc-428c-8f7a-25a1b36f16f9" containerName="extract-content" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.197352 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="680dc567-9fb8-4396-adb4-76bb44250c9c" containerName="registry-server" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.197361 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="98640d4a-0cbc-428c-8f7a-25a1b36f16f9" containerName="registry-server" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.197369 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d1c86c0-8a6f-4914-9593-20ca1220ed8b" containerName="registry-server" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.197376 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="337f295d-6c74-4850-bb14-125192de4385" containerName="marketplace-operator" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.197381 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="f59cbebb-100b-4d79-83f8-bf1a625caf93" containerName="registry-server" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.198050 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4zrs4" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.199838 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.205988 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4zrs4"] Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.250617 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfh7w\" (UniqueName: \"kubernetes.io/projected/b5a957ce-5a85-497c-9794-05ac9b6c92ae-kube-api-access-dfh7w\") pod \"redhat-marketplace-4zrs4\" (UID: \"b5a957ce-5a85-497c-9794-05ac9b6c92ae\") " pod="openshift-marketplace/redhat-marketplace-4zrs4" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.250672 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5a957ce-5a85-497c-9794-05ac9b6c92ae-utilities\") pod \"redhat-marketplace-4zrs4\" (UID: \"b5a957ce-5a85-497c-9794-05ac9b6c92ae\") " pod="openshift-marketplace/redhat-marketplace-4zrs4" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.250711 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5a957ce-5a85-497c-9794-05ac9b6c92ae-catalog-content\") pod \"redhat-marketplace-4zrs4\" (UID: \"b5a957ce-5a85-497c-9794-05ac9b6c92ae\") " pod="openshift-marketplace/redhat-marketplace-4zrs4" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.278009 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="337f295d-6c74-4850-bb14-125192de4385" path="/var/lib/kubelet/pods/337f295d-6c74-4850-bb14-125192de4385/volumes" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.278552 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="680dc567-9fb8-4396-adb4-76bb44250c9c" path="/var/lib/kubelet/pods/680dc567-9fb8-4396-adb4-76bb44250c9c/volumes" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.279312 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98640d4a-0cbc-428c-8f7a-25a1b36f16f9" path="/var/lib/kubelet/pods/98640d4a-0cbc-428c-8f7a-25a1b36f16f9/volumes" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.280223 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d1c86c0-8a6f-4914-9593-20ca1220ed8b" path="/var/lib/kubelet/pods/9d1c86c0-8a6f-4914-9593-20ca1220ed8b/volumes" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.280759 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f59cbebb-100b-4d79-83f8-bf1a625caf93" path="/var/lib/kubelet/pods/f59cbebb-100b-4d79-83f8-bf1a625caf93/volumes" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.351694 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfh7w\" (UniqueName: \"kubernetes.io/projected/b5a957ce-5a85-497c-9794-05ac9b6c92ae-kube-api-access-dfh7w\") pod \"redhat-marketplace-4zrs4\" (UID: \"b5a957ce-5a85-497c-9794-05ac9b6c92ae\") " pod="openshift-marketplace/redhat-marketplace-4zrs4" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.351732 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5a957ce-5a85-497c-9794-05ac9b6c92ae-utilities\") pod \"redhat-marketplace-4zrs4\" (UID: \"b5a957ce-5a85-497c-9794-05ac9b6c92ae\") " pod="openshift-marketplace/redhat-marketplace-4zrs4" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.351766 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5a957ce-5a85-497c-9794-05ac9b6c92ae-catalog-content\") pod \"redhat-marketplace-4zrs4\" (UID: \"b5a957ce-5a85-497c-9794-05ac9b6c92ae\") " pod="openshift-marketplace/redhat-marketplace-4zrs4" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.352328 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5a957ce-5a85-497c-9794-05ac9b6c92ae-catalog-content\") pod \"redhat-marketplace-4zrs4\" (UID: \"b5a957ce-5a85-497c-9794-05ac9b6c92ae\") " pod="openshift-marketplace/redhat-marketplace-4zrs4" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.352387 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5a957ce-5a85-497c-9794-05ac9b6c92ae-utilities\") pod \"redhat-marketplace-4zrs4\" (UID: \"b5a957ce-5a85-497c-9794-05ac9b6c92ae\") " pod="openshift-marketplace/redhat-marketplace-4zrs4" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.365802 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfh7w\" (UniqueName: \"kubernetes.io/projected/b5a957ce-5a85-497c-9794-05ac9b6c92ae-kube-api-access-dfh7w\") pod \"redhat-marketplace-4zrs4\" (UID: \"b5a957ce-5a85-497c-9794-05ac9b6c92ae\") " pod="openshift-marketplace/redhat-marketplace-4zrs4" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.514044 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.523770 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4zrs4" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.799764 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nc2rk"] Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.801725 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nc2rk" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.803670 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.805967 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nc2rk"] Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.857939 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-746qw\" (UniqueName: \"kubernetes.io/projected/4d8f5bd0-c208-43fb-acd6-496d94f6dec3-kube-api-access-746qw\") pod \"certified-operators-nc2rk\" (UID: \"4d8f5bd0-c208-43fb-acd6-496d94f6dec3\") " pod="openshift-marketplace/certified-operators-nc2rk" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.858055 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d8f5bd0-c208-43fb-acd6-496d94f6dec3-catalog-content\") pod \"certified-operators-nc2rk\" (UID: \"4d8f5bd0-c208-43fb-acd6-496d94f6dec3\") " pod="openshift-marketplace/certified-operators-nc2rk" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.858181 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d8f5bd0-c208-43fb-acd6-496d94f6dec3-utilities\") pod \"certified-operators-nc2rk\" (UID: \"4d8f5bd0-c208-43fb-acd6-496d94f6dec3\") " pod="openshift-marketplace/certified-operators-nc2rk" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.867945 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4zrs4"] Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.958877 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d8f5bd0-c208-43fb-acd6-496d94f6dec3-utilities\") pod \"certified-operators-nc2rk\" (UID: \"4d8f5bd0-c208-43fb-acd6-496d94f6dec3\") " pod="openshift-marketplace/certified-operators-nc2rk" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.958930 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-746qw\" (UniqueName: \"kubernetes.io/projected/4d8f5bd0-c208-43fb-acd6-496d94f6dec3-kube-api-access-746qw\") pod \"certified-operators-nc2rk\" (UID: \"4d8f5bd0-c208-43fb-acd6-496d94f6dec3\") " pod="openshift-marketplace/certified-operators-nc2rk" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.958985 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d8f5bd0-c208-43fb-acd6-496d94f6dec3-catalog-content\") pod \"certified-operators-nc2rk\" (UID: \"4d8f5bd0-c208-43fb-acd6-496d94f6dec3\") " pod="openshift-marketplace/certified-operators-nc2rk" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.959315 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d8f5bd0-c208-43fb-acd6-496d94f6dec3-utilities\") pod \"certified-operators-nc2rk\" (UID: \"4d8f5bd0-c208-43fb-acd6-496d94f6dec3\") " pod="openshift-marketplace/certified-operators-nc2rk" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.959387 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d8f5bd0-c208-43fb-acd6-496d94f6dec3-catalog-content\") pod \"certified-operators-nc2rk\" (UID: \"4d8f5bd0-c208-43fb-acd6-496d94f6dec3\") " pod="openshift-marketplace/certified-operators-nc2rk" Nov 23 06:48:58 crc kubenswrapper[4559]: I1123 06:48:58.973793 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-746qw\" (UniqueName: \"kubernetes.io/projected/4d8f5bd0-c208-43fb-acd6-496d94f6dec3-kube-api-access-746qw\") pod \"certified-operators-nc2rk\" (UID: \"4d8f5bd0-c208-43fb-acd6-496d94f6dec3\") " pod="openshift-marketplace/certified-operators-nc2rk" Nov 23 06:48:59 crc kubenswrapper[4559]: I1123 06:48:59.119583 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nc2rk" Nov 23 06:48:59 crc kubenswrapper[4559]: I1123 06:48:59.448487 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nc2rk"] Nov 23 06:48:59 crc kubenswrapper[4559]: W1123 06:48:59.452864 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d8f5bd0_c208_43fb_acd6_496d94f6dec3.slice/crio-0b19ab258b68641226fdd1a5ab6180d5e87dd7127ef2f6873afd2de0989ecd21 WatchSource:0}: Error finding container 0b19ab258b68641226fdd1a5ab6180d5e87dd7127ef2f6873afd2de0989ecd21: Status 404 returned error can't find the container with id 0b19ab258b68641226fdd1a5ab6180d5e87dd7127ef2f6873afd2de0989ecd21 Nov 23 06:48:59 crc kubenswrapper[4559]: I1123 06:48:59.642214 4559 generic.go:334] "Generic (PLEG): container finished" podID="b5a957ce-5a85-497c-9794-05ac9b6c92ae" containerID="b932ba595892247aded7aa3cb43e9e9d1608458f11cba074dc4ad6ee96449ad6" exitCode=0 Nov 23 06:48:59 crc kubenswrapper[4559]: I1123 06:48:59.642253 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4zrs4" event={"ID":"b5a957ce-5a85-497c-9794-05ac9b6c92ae","Type":"ContainerDied","Data":"b932ba595892247aded7aa3cb43e9e9d1608458f11cba074dc4ad6ee96449ad6"} Nov 23 06:48:59 crc kubenswrapper[4559]: I1123 06:48:59.642287 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4zrs4" event={"ID":"b5a957ce-5a85-497c-9794-05ac9b6c92ae","Type":"ContainerStarted","Data":"b114ba8ac652f3879f3f5dab317a7122d154b960fc2d9878ee4f132952089689"} Nov 23 06:48:59 crc kubenswrapper[4559]: I1123 06:48:59.644506 4559 generic.go:334] "Generic (PLEG): container finished" podID="4d8f5bd0-c208-43fb-acd6-496d94f6dec3" containerID="bf2eff80057b94aa18ae8c91a6f4765468dfe2b56de51a662335dbf89a8de577" exitCode=0 Nov 23 06:48:59 crc kubenswrapper[4559]: I1123 06:48:59.645322 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nc2rk" event={"ID":"4d8f5bd0-c208-43fb-acd6-496d94f6dec3","Type":"ContainerDied","Data":"bf2eff80057b94aa18ae8c91a6f4765468dfe2b56de51a662335dbf89a8de577"} Nov 23 06:48:59 crc kubenswrapper[4559]: I1123 06:48:59.645708 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nc2rk" event={"ID":"4d8f5bd0-c208-43fb-acd6-496d94f6dec3","Type":"ContainerStarted","Data":"0b19ab258b68641226fdd1a5ab6180d5e87dd7127ef2f6873afd2de0989ecd21"} Nov 23 06:49:00 crc kubenswrapper[4559]: I1123 06:49:00.596814 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w2mts"] Nov 23 06:49:00 crc kubenswrapper[4559]: I1123 06:49:00.599395 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w2mts" Nov 23 06:49:00 crc kubenswrapper[4559]: I1123 06:49:00.601993 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 23 06:49:00 crc kubenswrapper[4559]: I1123 06:49:00.604484 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w2mts"] Nov 23 06:49:00 crc kubenswrapper[4559]: I1123 06:49:00.649579 4559 generic.go:334] "Generic (PLEG): container finished" podID="b5a957ce-5a85-497c-9794-05ac9b6c92ae" containerID="8420a3cc9dedfd17436796305afe0d3fd0d849b053b888d367aeca5d1c74f92f" exitCode=0 Nov 23 06:49:00 crc kubenswrapper[4559]: I1123 06:49:00.649674 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4zrs4" event={"ID":"b5a957ce-5a85-497c-9794-05ac9b6c92ae","Type":"ContainerDied","Data":"8420a3cc9dedfd17436796305afe0d3fd0d849b053b888d367aeca5d1c74f92f"} Nov 23 06:49:00 crc kubenswrapper[4559]: I1123 06:49:00.652392 4559 generic.go:334] "Generic (PLEG): container finished" podID="4d8f5bd0-c208-43fb-acd6-496d94f6dec3" containerID="14db2a89864a84987bee5cb7a80e436446ce89d11ab58d6607e2179400ff0bac" exitCode=0 Nov 23 06:49:00 crc kubenswrapper[4559]: I1123 06:49:00.652422 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nc2rk" event={"ID":"4d8f5bd0-c208-43fb-acd6-496d94f6dec3","Type":"ContainerDied","Data":"14db2a89864a84987bee5cb7a80e436446ce89d11ab58d6607e2179400ff0bac"} Nov 23 06:49:00 crc kubenswrapper[4559]: I1123 06:49:00.673658 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6sc8\" (UniqueName: \"kubernetes.io/projected/0579529b-d4e5-458c-a2d6-9d0d7026c7b8-kube-api-access-n6sc8\") pod \"redhat-operators-w2mts\" (UID: \"0579529b-d4e5-458c-a2d6-9d0d7026c7b8\") " pod="openshift-marketplace/redhat-operators-w2mts" Nov 23 06:49:00 crc kubenswrapper[4559]: I1123 06:49:00.673745 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0579529b-d4e5-458c-a2d6-9d0d7026c7b8-catalog-content\") pod \"redhat-operators-w2mts\" (UID: \"0579529b-d4e5-458c-a2d6-9d0d7026c7b8\") " pod="openshift-marketplace/redhat-operators-w2mts" Nov 23 06:49:00 crc kubenswrapper[4559]: I1123 06:49:00.673784 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0579529b-d4e5-458c-a2d6-9d0d7026c7b8-utilities\") pod \"redhat-operators-w2mts\" (UID: \"0579529b-d4e5-458c-a2d6-9d0d7026c7b8\") " pod="openshift-marketplace/redhat-operators-w2mts" Nov 23 06:49:00 crc kubenswrapper[4559]: I1123 06:49:00.775084 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0579529b-d4e5-458c-a2d6-9d0d7026c7b8-utilities\") pod \"redhat-operators-w2mts\" (UID: \"0579529b-d4e5-458c-a2d6-9d0d7026c7b8\") " pod="openshift-marketplace/redhat-operators-w2mts" Nov 23 06:49:00 crc kubenswrapper[4559]: I1123 06:49:00.775128 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6sc8\" (UniqueName: \"kubernetes.io/projected/0579529b-d4e5-458c-a2d6-9d0d7026c7b8-kube-api-access-n6sc8\") pod \"redhat-operators-w2mts\" (UID: \"0579529b-d4e5-458c-a2d6-9d0d7026c7b8\") " pod="openshift-marketplace/redhat-operators-w2mts" Nov 23 06:49:00 crc kubenswrapper[4559]: I1123 06:49:00.775173 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0579529b-d4e5-458c-a2d6-9d0d7026c7b8-catalog-content\") pod \"redhat-operators-w2mts\" (UID: \"0579529b-d4e5-458c-a2d6-9d0d7026c7b8\") " pod="openshift-marketplace/redhat-operators-w2mts" Nov 23 06:49:00 crc kubenswrapper[4559]: I1123 06:49:00.775520 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0579529b-d4e5-458c-a2d6-9d0d7026c7b8-catalog-content\") pod \"redhat-operators-w2mts\" (UID: \"0579529b-d4e5-458c-a2d6-9d0d7026c7b8\") " pod="openshift-marketplace/redhat-operators-w2mts" Nov 23 06:49:00 crc kubenswrapper[4559]: I1123 06:49:00.775590 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0579529b-d4e5-458c-a2d6-9d0d7026c7b8-utilities\") pod \"redhat-operators-w2mts\" (UID: \"0579529b-d4e5-458c-a2d6-9d0d7026c7b8\") " pod="openshift-marketplace/redhat-operators-w2mts" Nov 23 06:49:00 crc kubenswrapper[4559]: I1123 06:49:00.790409 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6sc8\" (UniqueName: \"kubernetes.io/projected/0579529b-d4e5-458c-a2d6-9d0d7026c7b8-kube-api-access-n6sc8\") pod \"redhat-operators-w2mts\" (UID: \"0579529b-d4e5-458c-a2d6-9d0d7026c7b8\") " pod="openshift-marketplace/redhat-operators-w2mts" Nov 23 06:49:00 crc kubenswrapper[4559]: I1123 06:49:00.946405 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w2mts" Nov 23 06:49:01 crc kubenswrapper[4559]: I1123 06:49:01.199614 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6mnmc"] Nov 23 06:49:01 crc kubenswrapper[4559]: I1123 06:49:01.201884 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6mnmc" Nov 23 06:49:01 crc kubenswrapper[4559]: I1123 06:49:01.203585 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6mnmc"] Nov 23 06:49:01 crc kubenswrapper[4559]: I1123 06:49:01.204081 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 23 06:49:01 crc kubenswrapper[4559]: I1123 06:49:01.276442 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w2mts"] Nov 23 06:49:01 crc kubenswrapper[4559]: I1123 06:49:01.279479 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23ec849f-d94a-45b0-9f74-655f33ffb163-utilities\") pod \"community-operators-6mnmc\" (UID: \"23ec849f-d94a-45b0-9f74-655f33ffb163\") " pod="openshift-marketplace/community-operators-6mnmc" Nov 23 06:49:01 crc kubenswrapper[4559]: I1123 06:49:01.279530 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23ec849f-d94a-45b0-9f74-655f33ffb163-catalog-content\") pod \"community-operators-6mnmc\" (UID: \"23ec849f-d94a-45b0-9f74-655f33ffb163\") " pod="openshift-marketplace/community-operators-6mnmc" Nov 23 06:49:01 crc kubenswrapper[4559]: I1123 06:49:01.279566 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpgz5\" (UniqueName: \"kubernetes.io/projected/23ec849f-d94a-45b0-9f74-655f33ffb163-kube-api-access-hpgz5\") pod \"community-operators-6mnmc\" (UID: \"23ec849f-d94a-45b0-9f74-655f33ffb163\") " pod="openshift-marketplace/community-operators-6mnmc" Nov 23 06:49:01 crc kubenswrapper[4559]: I1123 06:49:01.380970 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23ec849f-d94a-45b0-9f74-655f33ffb163-utilities\") pod \"community-operators-6mnmc\" (UID: \"23ec849f-d94a-45b0-9f74-655f33ffb163\") " pod="openshift-marketplace/community-operators-6mnmc" Nov 23 06:49:01 crc kubenswrapper[4559]: I1123 06:49:01.381523 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23ec849f-d94a-45b0-9f74-655f33ffb163-catalog-content\") pod \"community-operators-6mnmc\" (UID: \"23ec849f-d94a-45b0-9f74-655f33ffb163\") " pod="openshift-marketplace/community-operators-6mnmc" Nov 23 06:49:01 crc kubenswrapper[4559]: I1123 06:49:01.381466 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23ec849f-d94a-45b0-9f74-655f33ffb163-utilities\") pod \"community-operators-6mnmc\" (UID: \"23ec849f-d94a-45b0-9f74-655f33ffb163\") " pod="openshift-marketplace/community-operators-6mnmc" Nov 23 06:49:01 crc kubenswrapper[4559]: I1123 06:49:01.381600 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpgz5\" (UniqueName: \"kubernetes.io/projected/23ec849f-d94a-45b0-9f74-655f33ffb163-kube-api-access-hpgz5\") pod \"community-operators-6mnmc\" (UID: \"23ec849f-d94a-45b0-9f74-655f33ffb163\") " pod="openshift-marketplace/community-operators-6mnmc" Nov 23 06:49:01 crc kubenswrapper[4559]: I1123 06:49:01.381978 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23ec849f-d94a-45b0-9f74-655f33ffb163-catalog-content\") pod \"community-operators-6mnmc\" (UID: \"23ec849f-d94a-45b0-9f74-655f33ffb163\") " pod="openshift-marketplace/community-operators-6mnmc" Nov 23 06:49:01 crc kubenswrapper[4559]: I1123 06:49:01.397905 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpgz5\" (UniqueName: \"kubernetes.io/projected/23ec849f-d94a-45b0-9f74-655f33ffb163-kube-api-access-hpgz5\") pod \"community-operators-6mnmc\" (UID: \"23ec849f-d94a-45b0-9f74-655f33ffb163\") " pod="openshift-marketplace/community-operators-6mnmc" Nov 23 06:49:01 crc kubenswrapper[4559]: I1123 06:49:01.526371 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6mnmc" Nov 23 06:49:01 crc kubenswrapper[4559]: I1123 06:49:01.659277 4559 generic.go:334] "Generic (PLEG): container finished" podID="0579529b-d4e5-458c-a2d6-9d0d7026c7b8" containerID="5f4fe88c913ac401ed5ec3b9389ce8c5bcf37471557040c0d18e4ed4535a58cd" exitCode=0 Nov 23 06:49:01 crc kubenswrapper[4559]: I1123 06:49:01.659332 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w2mts" event={"ID":"0579529b-d4e5-458c-a2d6-9d0d7026c7b8","Type":"ContainerDied","Data":"5f4fe88c913ac401ed5ec3b9389ce8c5bcf37471557040c0d18e4ed4535a58cd"} Nov 23 06:49:01 crc kubenswrapper[4559]: I1123 06:49:01.659528 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w2mts" event={"ID":"0579529b-d4e5-458c-a2d6-9d0d7026c7b8","Type":"ContainerStarted","Data":"422447b8156544b8054c30059fdc9b8b1547ad6ec091ae2132cca58f4b191952"} Nov 23 06:49:01 crc kubenswrapper[4559]: I1123 06:49:01.663316 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4zrs4" event={"ID":"b5a957ce-5a85-497c-9794-05ac9b6c92ae","Type":"ContainerStarted","Data":"55ad53066598c39513ab39f7b461769a63b5110a42a7da2818b8ddca7689c80b"} Nov 23 06:49:01 crc kubenswrapper[4559]: I1123 06:49:01.665190 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nc2rk" event={"ID":"4d8f5bd0-c208-43fb-acd6-496d94f6dec3","Type":"ContainerStarted","Data":"877c0b81d867c59cffca81de047c08ef05b8bde38abfc59744aed2269eb759a5"} Nov 23 06:49:01 crc kubenswrapper[4559]: I1123 06:49:01.684378 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4zrs4" podStartSLOduration=2.229373739 podStartE2EDuration="3.684363547s" podCreationTimestamp="2025-11-23 06:48:58 +0000 UTC" firstStartedPulling="2025-11-23 06:48:59.645070646 +0000 UTC m=+241.667056259" lastFinishedPulling="2025-11-23 06:49:01.100060453 +0000 UTC m=+243.122046067" observedRunningTime="2025-11-23 06:49:01.682845045 +0000 UTC m=+243.704830659" watchObservedRunningTime="2025-11-23 06:49:01.684363547 +0000 UTC m=+243.706349161" Nov 23 06:49:01 crc kubenswrapper[4559]: I1123 06:49:01.700785 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nc2rk" podStartSLOduration=2.17961623 podStartE2EDuration="3.700772228s" podCreationTimestamp="2025-11-23 06:48:58 +0000 UTC" firstStartedPulling="2025-11-23 06:48:59.645769257 +0000 UTC m=+241.667754872" lastFinishedPulling="2025-11-23 06:49:01.166925256 +0000 UTC m=+243.188910870" observedRunningTime="2025-11-23 06:49:01.700215703 +0000 UTC m=+243.722201317" watchObservedRunningTime="2025-11-23 06:49:01.700772228 +0000 UTC m=+243.722757842" Nov 23 06:49:01 crc kubenswrapper[4559]: I1123 06:49:01.878897 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6mnmc"] Nov 23 06:49:01 crc kubenswrapper[4559]: W1123 06:49:01.881152 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod23ec849f_d94a_45b0_9f74_655f33ffb163.slice/crio-16b68db09c6d1d2bbc74e64a3e61102d7bd957ae3993365edf0a977025e347dc WatchSource:0}: Error finding container 16b68db09c6d1d2bbc74e64a3e61102d7bd957ae3993365edf0a977025e347dc: Status 404 returned error can't find the container with id 16b68db09c6d1d2bbc74e64a3e61102d7bd957ae3993365edf0a977025e347dc Nov 23 06:49:02 crc kubenswrapper[4559]: I1123 06:49:02.669979 4559 generic.go:334] "Generic (PLEG): container finished" podID="23ec849f-d94a-45b0-9f74-655f33ffb163" containerID="d51204fb5c5393526d79c270481975d9d23198496963a689b1ed0ffbb3763c0e" exitCode=0 Nov 23 06:49:02 crc kubenswrapper[4559]: I1123 06:49:02.670078 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6mnmc" event={"ID":"23ec849f-d94a-45b0-9f74-655f33ffb163","Type":"ContainerDied","Data":"d51204fb5c5393526d79c270481975d9d23198496963a689b1ed0ffbb3763c0e"} Nov 23 06:49:02 crc kubenswrapper[4559]: I1123 06:49:02.670191 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6mnmc" event={"ID":"23ec849f-d94a-45b0-9f74-655f33ffb163","Type":"ContainerStarted","Data":"16b68db09c6d1d2bbc74e64a3e61102d7bd957ae3993365edf0a977025e347dc"} Nov 23 06:49:02 crc kubenswrapper[4559]: I1123 06:49:02.673473 4559 generic.go:334] "Generic (PLEG): container finished" podID="0579529b-d4e5-458c-a2d6-9d0d7026c7b8" containerID="20065e8df1e07b37bf0160a5840ba6051859b4f6bdfe3736639830fb5dd87cb0" exitCode=0 Nov 23 06:49:02 crc kubenswrapper[4559]: I1123 06:49:02.673607 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w2mts" event={"ID":"0579529b-d4e5-458c-a2d6-9d0d7026c7b8","Type":"ContainerDied","Data":"20065e8df1e07b37bf0160a5840ba6051859b4f6bdfe3736639830fb5dd87cb0"} Nov 23 06:49:04 crc kubenswrapper[4559]: I1123 06:49:04.681345 4559 generic.go:334] "Generic (PLEG): container finished" podID="23ec849f-d94a-45b0-9f74-655f33ffb163" containerID="1973e1efebe3c33a41615a39f1f0a95af29d8087ad7606cfac7e95d9f0fda64f" exitCode=0 Nov 23 06:49:04 crc kubenswrapper[4559]: I1123 06:49:04.681428 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6mnmc" event={"ID":"23ec849f-d94a-45b0-9f74-655f33ffb163","Type":"ContainerDied","Data":"1973e1efebe3c33a41615a39f1f0a95af29d8087ad7606cfac7e95d9f0fda64f"} Nov 23 06:49:04 crc kubenswrapper[4559]: I1123 06:49:04.685263 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w2mts" event={"ID":"0579529b-d4e5-458c-a2d6-9d0d7026c7b8","Type":"ContainerStarted","Data":"eb0e9be969e3fa1dfde02876801c9ae56e8cab4167bfbd1b11b3c94458f5ed79"} Nov 23 06:49:04 crc kubenswrapper[4559]: I1123 06:49:04.714200 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w2mts" podStartSLOduration=3.198426122 podStartE2EDuration="4.714188238s" podCreationTimestamp="2025-11-23 06:49:00 +0000 UTC" firstStartedPulling="2025-11-23 06:49:01.660706351 +0000 UTC m=+243.682691966" lastFinishedPulling="2025-11-23 06:49:03.176468467 +0000 UTC m=+245.198454082" observedRunningTime="2025-11-23 06:49:04.713528228 +0000 UTC m=+246.735513852" watchObservedRunningTime="2025-11-23 06:49:04.714188238 +0000 UTC m=+246.736173851" Nov 23 06:49:05 crc kubenswrapper[4559]: I1123 06:49:05.691597 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6mnmc" event={"ID":"23ec849f-d94a-45b0-9f74-655f33ffb163","Type":"ContainerStarted","Data":"6b00c16d84a1cad97b06f59023b6089017c02695da06a30ef74711818044c86e"} Nov 23 06:49:05 crc kubenswrapper[4559]: I1123 06:49:05.704024 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6mnmc" podStartSLOduration=2.223208221 podStartE2EDuration="4.704012498s" podCreationTimestamp="2025-11-23 06:49:01 +0000 UTC" firstStartedPulling="2025-11-23 06:49:02.67138867 +0000 UTC m=+244.693374284" lastFinishedPulling="2025-11-23 06:49:05.152192948 +0000 UTC m=+247.174178561" observedRunningTime="2025-11-23 06:49:05.702925236 +0000 UTC m=+247.724910849" watchObservedRunningTime="2025-11-23 06:49:05.704012498 +0000 UTC m=+247.725998112" Nov 23 06:49:08 crc kubenswrapper[4559]: I1123 06:49:08.524678 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4zrs4" Nov 23 06:49:08 crc kubenswrapper[4559]: I1123 06:49:08.524911 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4zrs4" Nov 23 06:49:08 crc kubenswrapper[4559]: I1123 06:49:08.551725 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4zrs4" Nov 23 06:49:08 crc kubenswrapper[4559]: I1123 06:49:08.726289 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4zrs4" Nov 23 06:49:09 crc kubenswrapper[4559]: I1123 06:49:09.119993 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nc2rk" Nov 23 06:49:09 crc kubenswrapper[4559]: I1123 06:49:09.120030 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nc2rk" Nov 23 06:49:09 crc kubenswrapper[4559]: I1123 06:49:09.145747 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nc2rk" Nov 23 06:49:09 crc kubenswrapper[4559]: I1123 06:49:09.729166 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nc2rk" Nov 23 06:49:10 crc kubenswrapper[4559]: I1123 06:49:10.947679 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-w2mts" Nov 23 06:49:10 crc kubenswrapper[4559]: I1123 06:49:10.947717 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-w2mts" Nov 23 06:49:10 crc kubenswrapper[4559]: I1123 06:49:10.972542 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-w2mts" Nov 23 06:49:11 crc kubenswrapper[4559]: I1123 06:49:11.526966 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6mnmc" Nov 23 06:49:11 crc kubenswrapper[4559]: I1123 06:49:11.527015 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6mnmc" Nov 23 06:49:11 crc kubenswrapper[4559]: I1123 06:49:11.554795 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6mnmc" Nov 23 06:49:11 crc kubenswrapper[4559]: I1123 06:49:11.745821 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-w2mts" Nov 23 06:49:11 crc kubenswrapper[4559]: I1123 06:49:11.762083 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6mnmc" Nov 23 06:50:26 crc kubenswrapper[4559]: I1123 06:50:26.167392 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 06:50:26 crc kubenswrapper[4559]: I1123 06:50:26.167789 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 06:50:56 crc kubenswrapper[4559]: I1123 06:50:56.167332 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 06:50:56 crc kubenswrapper[4559]: I1123 06:50:56.167708 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 06:51:26 crc kubenswrapper[4559]: I1123 06:51:26.166683 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 06:51:26 crc kubenswrapper[4559]: I1123 06:51:26.167034 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 06:51:26 crc kubenswrapper[4559]: I1123 06:51:26.167067 4559 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" Nov 23 06:51:26 crc kubenswrapper[4559]: I1123 06:51:26.167428 4559 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"232a78e72da23760d36189f8c69ea7f78803b94ef0a735cc1b9c323800c5683a"} pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 06:51:26 crc kubenswrapper[4559]: I1123 06:51:26.167474 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" containerID="cri-o://232a78e72da23760d36189f8c69ea7f78803b94ef0a735cc1b9c323800c5683a" gracePeriod=600 Nov 23 06:51:26 crc kubenswrapper[4559]: I1123 06:51:26.299837 4559 generic.go:334] "Generic (PLEG): container finished" podID="4731beee-0cac-4189-8a70-743b0b709095" containerID="232a78e72da23760d36189f8c69ea7f78803b94ef0a735cc1b9c323800c5683a" exitCode=0 Nov 23 06:51:26 crc kubenswrapper[4559]: I1123 06:51:26.299866 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerDied","Data":"232a78e72da23760d36189f8c69ea7f78803b94ef0a735cc1b9c323800c5683a"} Nov 23 06:51:26 crc kubenswrapper[4559]: I1123 06:51:26.299889 4559 scope.go:117] "RemoveContainer" containerID="053843214b80ee8e871c4eb2d303f81637b5f52e28452c6cbac4f57ab1424f43" Nov 23 06:51:27 crc kubenswrapper[4559]: I1123 06:51:27.304982 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerStarted","Data":"caa8c42bf6f598aae815f53f4281cd35582267034e69906ef9d6b656540ea6ab"} Nov 23 06:52:49 crc kubenswrapper[4559]: I1123 06:52:49.887837 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-7dhmb"] Nov 23 06:52:49 crc kubenswrapper[4559]: I1123 06:52:49.888714 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" Nov 23 06:52:49 crc kubenswrapper[4559]: I1123 06:52:49.898352 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-7dhmb"] Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.043327 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hvkg\" (UniqueName: \"kubernetes.io/projected/639552be-fe3e-4114-967d-76a5508534a3-kube-api-access-2hvkg\") pod \"image-registry-66df7c8f76-7dhmb\" (UID: \"639552be-fe3e-4114-967d-76a5508534a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.043375 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/639552be-fe3e-4114-967d-76a5508534a3-bound-sa-token\") pod \"image-registry-66df7c8f76-7dhmb\" (UID: \"639552be-fe3e-4114-967d-76a5508534a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.043439 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/639552be-fe3e-4114-967d-76a5508534a3-registry-certificates\") pod \"image-registry-66df7c8f76-7dhmb\" (UID: \"639552be-fe3e-4114-967d-76a5508534a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.043467 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/639552be-fe3e-4114-967d-76a5508534a3-registry-tls\") pod \"image-registry-66df7c8f76-7dhmb\" (UID: \"639552be-fe3e-4114-967d-76a5508534a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.043486 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/639552be-fe3e-4114-967d-76a5508534a3-trusted-ca\") pod \"image-registry-66df7c8f76-7dhmb\" (UID: \"639552be-fe3e-4114-967d-76a5508534a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.043522 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-7dhmb\" (UID: \"639552be-fe3e-4114-967d-76a5508534a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.043544 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/639552be-fe3e-4114-967d-76a5508534a3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-7dhmb\" (UID: \"639552be-fe3e-4114-967d-76a5508534a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.043624 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/639552be-fe3e-4114-967d-76a5508534a3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-7dhmb\" (UID: \"639552be-fe3e-4114-967d-76a5508534a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.061863 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-7dhmb\" (UID: \"639552be-fe3e-4114-967d-76a5508534a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.144476 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/639552be-fe3e-4114-967d-76a5508534a3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-7dhmb\" (UID: \"639552be-fe3e-4114-967d-76a5508534a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.144553 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hvkg\" (UniqueName: \"kubernetes.io/projected/639552be-fe3e-4114-967d-76a5508534a3-kube-api-access-2hvkg\") pod \"image-registry-66df7c8f76-7dhmb\" (UID: \"639552be-fe3e-4114-967d-76a5508534a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.144582 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/639552be-fe3e-4114-967d-76a5508534a3-bound-sa-token\") pod \"image-registry-66df7c8f76-7dhmb\" (UID: \"639552be-fe3e-4114-967d-76a5508534a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.144623 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/639552be-fe3e-4114-967d-76a5508534a3-registry-tls\") pod \"image-registry-66df7c8f76-7dhmb\" (UID: \"639552be-fe3e-4114-967d-76a5508534a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.144685 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/639552be-fe3e-4114-967d-76a5508534a3-registry-certificates\") pod \"image-registry-66df7c8f76-7dhmb\" (UID: \"639552be-fe3e-4114-967d-76a5508534a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.144714 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/639552be-fe3e-4114-967d-76a5508534a3-trusted-ca\") pod \"image-registry-66df7c8f76-7dhmb\" (UID: \"639552be-fe3e-4114-967d-76a5508534a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.144763 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/639552be-fe3e-4114-967d-76a5508534a3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-7dhmb\" (UID: \"639552be-fe3e-4114-967d-76a5508534a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.144969 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/639552be-fe3e-4114-967d-76a5508534a3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-7dhmb\" (UID: \"639552be-fe3e-4114-967d-76a5508534a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.146041 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/639552be-fe3e-4114-967d-76a5508534a3-trusted-ca\") pod \"image-registry-66df7c8f76-7dhmb\" (UID: \"639552be-fe3e-4114-967d-76a5508534a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.146363 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/639552be-fe3e-4114-967d-76a5508534a3-registry-certificates\") pod \"image-registry-66df7c8f76-7dhmb\" (UID: \"639552be-fe3e-4114-967d-76a5508534a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.149030 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/639552be-fe3e-4114-967d-76a5508534a3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-7dhmb\" (UID: \"639552be-fe3e-4114-967d-76a5508534a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.149499 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/639552be-fe3e-4114-967d-76a5508534a3-registry-tls\") pod \"image-registry-66df7c8f76-7dhmb\" (UID: \"639552be-fe3e-4114-967d-76a5508534a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.157160 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/639552be-fe3e-4114-967d-76a5508534a3-bound-sa-token\") pod \"image-registry-66df7c8f76-7dhmb\" (UID: \"639552be-fe3e-4114-967d-76a5508534a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.157429 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hvkg\" (UniqueName: \"kubernetes.io/projected/639552be-fe3e-4114-967d-76a5508534a3-kube-api-access-2hvkg\") pod \"image-registry-66df7c8f76-7dhmb\" (UID: \"639552be-fe3e-4114-967d-76a5508534a3\") " pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.201409 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.355328 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-7dhmb"] Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.596719 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" event={"ID":"639552be-fe3e-4114-967d-76a5508534a3","Type":"ContainerStarted","Data":"0f84449a24183ae4090bac091559d55dd14b3f5524b67c072ebc783eae75e7cc"} Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.596771 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" event={"ID":"639552be-fe3e-4114-967d-76a5508534a3","Type":"ContainerStarted","Data":"fdab1f5ead182e52462fe2a0f9f0475d0b620665389b8fe07a596c0f03284ba7"} Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.596887 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" Nov 23 06:52:50 crc kubenswrapper[4559]: I1123 06:52:50.611859 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" podStartSLOduration=1.611840676 podStartE2EDuration="1.611840676s" podCreationTimestamp="2025-11-23 06:52:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:52:50.60928664 +0000 UTC m=+472.631272254" watchObservedRunningTime="2025-11-23 06:52:50.611840676 +0000 UTC m=+472.633826291" Nov 23 06:53:10 crc kubenswrapper[4559]: I1123 06:53:10.204982 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-7dhmb" Nov 23 06:53:10 crc kubenswrapper[4559]: I1123 06:53:10.239995 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ccrq2"] Nov 23 06:53:26 crc kubenswrapper[4559]: I1123 06:53:26.166732 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 06:53:26 crc kubenswrapper[4559]: I1123 06:53:26.167635 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 06:53:27 crc kubenswrapper[4559]: I1123 06:53:27.570011 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-m6h87"] Nov 23 06:53:27 crc kubenswrapper[4559]: I1123 06:53:27.570823 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-m6h87" Nov 23 06:53:27 crc kubenswrapper[4559]: I1123 06:53:27.572473 4559 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-h5dk6" Nov 23 06:53:27 crc kubenswrapper[4559]: I1123 06:53:27.572709 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 23 06:53:27 crc kubenswrapper[4559]: I1123 06:53:27.573914 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 23 06:53:27 crc kubenswrapper[4559]: I1123 06:53:27.581231 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-m6h87"] Nov 23 06:53:27 crc kubenswrapper[4559]: I1123 06:53:27.587552 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-mqxbb"] Nov 23 06:53:27 crc kubenswrapper[4559]: I1123 06:53:27.588079 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-mqxbb" Nov 23 06:53:27 crc kubenswrapper[4559]: I1123 06:53:27.590273 4559 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-gc2px" Nov 23 06:53:27 crc kubenswrapper[4559]: I1123 06:53:27.590579 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-qxxh5"] Nov 23 06:53:27 crc kubenswrapper[4559]: I1123 06:53:27.591161 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-qxxh5" Nov 23 06:53:27 crc kubenswrapper[4559]: I1123 06:53:27.596599 4559 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-kvlr4" Nov 23 06:53:27 crc kubenswrapper[4559]: I1123 06:53:27.599042 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-mqxbb"] Nov 23 06:53:27 crc kubenswrapper[4559]: I1123 06:53:27.601658 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-qxxh5"] Nov 23 06:53:27 crc kubenswrapper[4559]: I1123 06:53:27.761180 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gn4r2\" (UniqueName: \"kubernetes.io/projected/74a9e3f9-a1b2-4f28-9ff4-a6d8a5d793b7-kube-api-access-gn4r2\") pod \"cert-manager-webhook-5655c58dd6-qxxh5\" (UID: \"74a9e3f9-a1b2-4f28-9ff4-a6d8a5d793b7\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-qxxh5" Nov 23 06:53:27 crc kubenswrapper[4559]: I1123 06:53:27.761257 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvdnf\" (UniqueName: \"kubernetes.io/projected/eb252f53-26dd-4071-8010-8a824926a679-kube-api-access-zvdnf\") pod \"cert-manager-cainjector-7f985d654d-m6h87\" (UID: \"eb252f53-26dd-4071-8010-8a824926a679\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-m6h87" Nov 23 06:53:27 crc kubenswrapper[4559]: I1123 06:53:27.761282 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xg25s\" (UniqueName: \"kubernetes.io/projected/e5b5c7a9-2ccb-44ac-a6bd-826ffa8d7bda-kube-api-access-xg25s\") pod \"cert-manager-5b446d88c5-mqxbb\" (UID: \"e5b5c7a9-2ccb-44ac-a6bd-826ffa8d7bda\") " pod="cert-manager/cert-manager-5b446d88c5-mqxbb" Nov 23 06:53:27 crc kubenswrapper[4559]: I1123 06:53:27.862268 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gn4r2\" (UniqueName: \"kubernetes.io/projected/74a9e3f9-a1b2-4f28-9ff4-a6d8a5d793b7-kube-api-access-gn4r2\") pod \"cert-manager-webhook-5655c58dd6-qxxh5\" (UID: \"74a9e3f9-a1b2-4f28-9ff4-a6d8a5d793b7\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-qxxh5" Nov 23 06:53:27 crc kubenswrapper[4559]: I1123 06:53:27.862522 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvdnf\" (UniqueName: \"kubernetes.io/projected/eb252f53-26dd-4071-8010-8a824926a679-kube-api-access-zvdnf\") pod \"cert-manager-cainjector-7f985d654d-m6h87\" (UID: \"eb252f53-26dd-4071-8010-8a824926a679\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-m6h87" Nov 23 06:53:27 crc kubenswrapper[4559]: I1123 06:53:27.862615 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xg25s\" (UniqueName: \"kubernetes.io/projected/e5b5c7a9-2ccb-44ac-a6bd-826ffa8d7bda-kube-api-access-xg25s\") pod \"cert-manager-5b446d88c5-mqxbb\" (UID: \"e5b5c7a9-2ccb-44ac-a6bd-826ffa8d7bda\") " pod="cert-manager/cert-manager-5b446d88c5-mqxbb" Nov 23 06:53:27 crc kubenswrapper[4559]: I1123 06:53:27.876413 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gn4r2\" (UniqueName: \"kubernetes.io/projected/74a9e3f9-a1b2-4f28-9ff4-a6d8a5d793b7-kube-api-access-gn4r2\") pod \"cert-manager-webhook-5655c58dd6-qxxh5\" (UID: \"74a9e3f9-a1b2-4f28-9ff4-a6d8a5d793b7\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-qxxh5" Nov 23 06:53:27 crc kubenswrapper[4559]: I1123 06:53:27.876427 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xg25s\" (UniqueName: \"kubernetes.io/projected/e5b5c7a9-2ccb-44ac-a6bd-826ffa8d7bda-kube-api-access-xg25s\") pod \"cert-manager-5b446d88c5-mqxbb\" (UID: \"e5b5c7a9-2ccb-44ac-a6bd-826ffa8d7bda\") " pod="cert-manager/cert-manager-5b446d88c5-mqxbb" Nov 23 06:53:27 crc kubenswrapper[4559]: I1123 06:53:27.876673 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvdnf\" (UniqueName: \"kubernetes.io/projected/eb252f53-26dd-4071-8010-8a824926a679-kube-api-access-zvdnf\") pod \"cert-manager-cainjector-7f985d654d-m6h87\" (UID: \"eb252f53-26dd-4071-8010-8a824926a679\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-m6h87" Nov 23 06:53:27 crc kubenswrapper[4559]: I1123 06:53:27.886197 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-m6h87" Nov 23 06:53:27 crc kubenswrapper[4559]: I1123 06:53:27.901063 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-mqxbb" Nov 23 06:53:27 crc kubenswrapper[4559]: I1123 06:53:27.905982 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-qxxh5" Nov 23 06:53:28 crc kubenswrapper[4559]: I1123 06:53:28.222256 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-m6h87"] Nov 23 06:53:28 crc kubenswrapper[4559]: I1123 06:53:28.227693 4559 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 06:53:28 crc kubenswrapper[4559]: I1123 06:53:28.259411 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-qxxh5"] Nov 23 06:53:28 crc kubenswrapper[4559]: W1123 06:53:28.261100 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74a9e3f9_a1b2_4f28_9ff4_a6d8a5d793b7.slice/crio-b52fd5ed6e0155408d3614d9843465f7131f0723364969883d12fed93e0f27cc WatchSource:0}: Error finding container b52fd5ed6e0155408d3614d9843465f7131f0723364969883d12fed93e0f27cc: Status 404 returned error can't find the container with id b52fd5ed6e0155408d3614d9843465f7131f0723364969883d12fed93e0f27cc Nov 23 06:53:28 crc kubenswrapper[4559]: I1123 06:53:28.261365 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-mqxbb"] Nov 23 06:53:28 crc kubenswrapper[4559]: W1123 06:53:28.266122 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5b5c7a9_2ccb_44ac_a6bd_826ffa8d7bda.slice/crio-d712d290fb51befad5ec5b72181bf97263ceea58f69c35388fedde78a6e027d0 WatchSource:0}: Error finding container d712d290fb51befad5ec5b72181bf97263ceea58f69c35388fedde78a6e027d0: Status 404 returned error can't find the container with id d712d290fb51befad5ec5b72181bf97263ceea58f69c35388fedde78a6e027d0 Nov 23 06:53:28 crc kubenswrapper[4559]: I1123 06:53:28.755826 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-m6h87" event={"ID":"eb252f53-26dd-4071-8010-8a824926a679","Type":"ContainerStarted","Data":"494218d993cb7d04ac49bd5b7db4e1ae6f38e6e567b1d7888a11114f9c0e0310"} Nov 23 06:53:28 crc kubenswrapper[4559]: I1123 06:53:28.757205 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-mqxbb" event={"ID":"e5b5c7a9-2ccb-44ac-a6bd-826ffa8d7bda","Type":"ContainerStarted","Data":"d712d290fb51befad5ec5b72181bf97263ceea58f69c35388fedde78a6e027d0"} Nov 23 06:53:28 crc kubenswrapper[4559]: I1123 06:53:28.758097 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-qxxh5" event={"ID":"74a9e3f9-a1b2-4f28-9ff4-a6d8a5d793b7","Type":"ContainerStarted","Data":"b52fd5ed6e0155408d3614d9843465f7131f0723364969883d12fed93e0f27cc"} Nov 23 06:53:30 crc kubenswrapper[4559]: I1123 06:53:30.768573 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-m6h87" event={"ID":"eb252f53-26dd-4071-8010-8a824926a679","Type":"ContainerStarted","Data":"900a088551dfbb4811c592b177f3d4119380d221a1a9613de424f04aca714eef"} Nov 23 06:53:30 crc kubenswrapper[4559]: I1123 06:53:30.771514 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-qxxh5" event={"ID":"74a9e3f9-a1b2-4f28-9ff4-a6d8a5d793b7","Type":"ContainerStarted","Data":"b11fd15abbf7ec7d8b53bdc5e07ebf6dfdacc0ad01ffb50d1fc57a7bc1a375e3"} Nov 23 06:53:30 crc kubenswrapper[4559]: I1123 06:53:30.771729 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-qxxh5" Nov 23 06:53:30 crc kubenswrapper[4559]: I1123 06:53:30.772739 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-mqxbb" event={"ID":"e5b5c7a9-2ccb-44ac-a6bd-826ffa8d7bda","Type":"ContainerStarted","Data":"252db410787c6c7fadbe0447c0cca0d7742abeab29f393251ce22735f7fb2e1e"} Nov 23 06:53:30 crc kubenswrapper[4559]: I1123 06:53:30.795348 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-mqxbb" podStartSLOduration=1.4938301410000001 podStartE2EDuration="3.795328338s" podCreationTimestamp="2025-11-23 06:53:27 +0000 UTC" firstStartedPulling="2025-11-23 06:53:28.267742989 +0000 UTC m=+510.289728603" lastFinishedPulling="2025-11-23 06:53:30.569241195 +0000 UTC m=+512.591226800" observedRunningTime="2025-11-23 06:53:30.794047598 +0000 UTC m=+512.816033213" watchObservedRunningTime="2025-11-23 06:53:30.795328338 +0000 UTC m=+512.817313952" Nov 23 06:53:30 crc kubenswrapper[4559]: I1123 06:53:30.796026 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-m6h87" podStartSLOduration=1.564524668 podStartE2EDuration="3.796018637s" podCreationTimestamp="2025-11-23 06:53:27 +0000 UTC" firstStartedPulling="2025-11-23 06:53:28.227439134 +0000 UTC m=+510.249424748" lastFinishedPulling="2025-11-23 06:53:30.458933103 +0000 UTC m=+512.480918717" observedRunningTime="2025-11-23 06:53:30.780761506 +0000 UTC m=+512.802747120" watchObservedRunningTime="2025-11-23 06:53:30.796018637 +0000 UTC m=+512.818004251" Nov 23 06:53:30 crc kubenswrapper[4559]: I1123 06:53:30.815091 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-qxxh5" podStartSLOduration=1.533655266 podStartE2EDuration="3.81507578s" podCreationTimestamp="2025-11-23 06:53:27 +0000 UTC" firstStartedPulling="2025-11-23 06:53:28.263151669 +0000 UTC m=+510.285137283" lastFinishedPulling="2025-11-23 06:53:30.544572183 +0000 UTC m=+512.566557797" observedRunningTime="2025-11-23 06:53:30.806774669 +0000 UTC m=+512.828760283" watchObservedRunningTime="2025-11-23 06:53:30.81507578 +0000 UTC m=+512.837061394" Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.266219 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" podUID="2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d" containerName="registry" containerID="cri-o://4756d4be1b1c5e4807c762f8492154b9a2f27691c5d2c91d3f70165db42be5d3" gracePeriod=30 Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.531558 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.643821 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-ca-trust-extracted\") pod \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.643856 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-installation-pull-secrets\") pod \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.643904 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-registry-certificates\") pod \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.643931 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-trusted-ca\") pod \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.643954 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-bound-sa-token\") pod \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.643985 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-registry-tls\") pod \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.644009 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rg7c6\" (UniqueName: \"kubernetes.io/projected/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-kube-api-access-rg7c6\") pod \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.644142 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\" (UID: \"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d\") " Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.644756 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.645306 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.649148 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.649250 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.649352 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-kube-api-access-rg7c6" (OuterVolumeSpecName: "kube-api-access-rg7c6") pod "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d"). InnerVolumeSpecName "kube-api-access-rg7c6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.649506 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.651289 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.658016 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d" (UID: "2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.745681 4559 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.745859 4559 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.745871 4559 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.745880 4559 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.745889 4559 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.745898 4559 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.745906 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rg7c6\" (UniqueName: \"kubernetes.io/projected/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d-kube-api-access-rg7c6\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.795749 4559 generic.go:334] "Generic (PLEG): container finished" podID="2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d" containerID="4756d4be1b1c5e4807c762f8492154b9a2f27691c5d2c91d3f70165db42be5d3" exitCode=0 Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.795800 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.795788 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" event={"ID":"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d","Type":"ContainerDied","Data":"4756d4be1b1c5e4807c762f8492154b9a2f27691c5d2c91d3f70165db42be5d3"} Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.795924 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-ccrq2" event={"ID":"2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d","Type":"ContainerDied","Data":"839794abcb333c8ab2d3709f630533953507fcd49b8f9a3eb3c80da034b03694"} Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.795944 4559 scope.go:117] "RemoveContainer" containerID="4756d4be1b1c5e4807c762f8492154b9a2f27691c5d2c91d3f70165db42be5d3" Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.808618 4559 scope.go:117] "RemoveContainer" containerID="4756d4be1b1c5e4807c762f8492154b9a2f27691c5d2c91d3f70165db42be5d3" Nov 23 06:53:35 crc kubenswrapper[4559]: E1123 06:53:35.808913 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4756d4be1b1c5e4807c762f8492154b9a2f27691c5d2c91d3f70165db42be5d3\": container with ID starting with 4756d4be1b1c5e4807c762f8492154b9a2f27691c5d2c91d3f70165db42be5d3 not found: ID does not exist" containerID="4756d4be1b1c5e4807c762f8492154b9a2f27691c5d2c91d3f70165db42be5d3" Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.808944 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4756d4be1b1c5e4807c762f8492154b9a2f27691c5d2c91d3f70165db42be5d3"} err="failed to get container status \"4756d4be1b1c5e4807c762f8492154b9a2f27691c5d2c91d3f70165db42be5d3\": rpc error: code = NotFound desc = could not find container \"4756d4be1b1c5e4807c762f8492154b9a2f27691c5d2c91d3f70165db42be5d3\": container with ID starting with 4756d4be1b1c5e4807c762f8492154b9a2f27691c5d2c91d3f70165db42be5d3 not found: ID does not exist" Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.815612 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ccrq2"] Nov 23 06:53:35 crc kubenswrapper[4559]: I1123 06:53:35.818129 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ccrq2"] Nov 23 06:53:36 crc kubenswrapper[4559]: I1123 06:53:36.278034 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d" path="/var/lib/kubelet/pods/2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d/volumes" Nov 23 06:53:37 crc kubenswrapper[4559]: I1123 06:53:37.909518 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-qxxh5" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.319306 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jt56f"] Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.319618 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="ovn-controller" containerID="cri-o://96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb" gracePeriod=30 Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.319688 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="nbdb" containerID="cri-o://1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb" gracePeriod=30 Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.319762 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="northd" containerID="cri-o://664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3" gracePeriod=30 Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.319779 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="sbdb" containerID="cri-o://52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90" gracePeriod=30 Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.319801 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d" gracePeriod=30 Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.319833 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="kube-rbac-proxy-node" containerID="cri-o://ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77" gracePeriod=30 Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.319857 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="ovn-acl-logging" containerID="cri-o://651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565" gracePeriod=30 Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.360149 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="ovnkube-controller" containerID="cri-o://0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8" gracePeriod=30 Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.577047 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jt56f_b12e257e-7887-4795-9221-4db8fd5856dd/ovnkube-controller/3.log" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.578913 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jt56f_b12e257e-7887-4795-9221-4db8fd5856dd/ovn-acl-logging/0.log" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.579314 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jt56f_b12e257e-7887-4795-9221-4db8fd5856dd/ovn-controller/0.log" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.579691 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.617794 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lv845"] Nov 23 06:53:39 crc kubenswrapper[4559]: E1123 06:53:39.617971 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="ovnkube-controller" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.617987 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="ovnkube-controller" Nov 23 06:53:39 crc kubenswrapper[4559]: E1123 06:53:39.617995 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="ovn-acl-logging" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.618001 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="ovn-acl-logging" Nov 23 06:53:39 crc kubenswrapper[4559]: E1123 06:53:39.618009 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="ovnkube-controller" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.618014 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="ovnkube-controller" Nov 23 06:53:39 crc kubenswrapper[4559]: E1123 06:53:39.618020 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="kube-rbac-proxy-node" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.618025 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="kube-rbac-proxy-node" Nov 23 06:53:39 crc kubenswrapper[4559]: E1123 06:53:39.618032 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="ovnkube-controller" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.618037 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="ovnkube-controller" Nov 23 06:53:39 crc kubenswrapper[4559]: E1123 06:53:39.618043 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="kube-rbac-proxy-ovn-metrics" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.618049 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="kube-rbac-proxy-ovn-metrics" Nov 23 06:53:39 crc kubenswrapper[4559]: E1123 06:53:39.618057 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="northd" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.618062 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="northd" Nov 23 06:53:39 crc kubenswrapper[4559]: E1123 06:53:39.618069 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="kubecfg-setup" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.618074 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="kubecfg-setup" Nov 23 06:53:39 crc kubenswrapper[4559]: E1123 06:53:39.618082 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="ovn-controller" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.618087 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="ovn-controller" Nov 23 06:53:39 crc kubenswrapper[4559]: E1123 06:53:39.618095 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d" containerName="registry" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.618100 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d" containerName="registry" Nov 23 06:53:39 crc kubenswrapper[4559]: E1123 06:53:39.618105 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="nbdb" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.618110 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="nbdb" Nov 23 06:53:39 crc kubenswrapper[4559]: E1123 06:53:39.618118 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="sbdb" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.618122 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="sbdb" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.618201 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="ovnkube-controller" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.618212 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="kube-rbac-proxy-ovn-metrics" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.618218 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aee8d30-e0ce-4b6a-8e3c-d5e42d0bf87d" containerName="registry" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.618224 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="ovnkube-controller" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.618230 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="ovnkube-controller" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.618237 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="northd" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.618242 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="sbdb" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.618250 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="nbdb" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.618256 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="ovn-acl-logging" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.618262 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="ovnkube-controller" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.618269 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="ovn-controller" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.618275 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="kube-rbac-proxy-node" Nov 23 06:53:39 crc kubenswrapper[4559]: E1123 06:53:39.618349 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="ovnkube-controller" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.618356 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="ovnkube-controller" Nov 23 06:53:39 crc kubenswrapper[4559]: E1123 06:53:39.618365 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="ovnkube-controller" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.618370 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="ovnkube-controller" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.618445 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" containerName="ovnkube-controller" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.619725 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.682153 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-cni-netd\") pod \"b12e257e-7887-4795-9221-4db8fd5856dd\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.682187 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-run-netns\") pod \"b12e257e-7887-4795-9221-4db8fd5856dd\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.682213 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b12e257e-7887-4795-9221-4db8fd5856dd-ovnkube-config\") pod \"b12e257e-7887-4795-9221-4db8fd5856dd\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.682238 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-kubelet\") pod \"b12e257e-7887-4795-9221-4db8fd5856dd\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.682259 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b12e257e-7887-4795-9221-4db8fd5856dd-ovn-node-metrics-cert\") pod \"b12e257e-7887-4795-9221-4db8fd5856dd\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.682299 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-systemd-units\") pod \"b12e257e-7887-4795-9221-4db8fd5856dd\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.682312 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-slash\") pod \"b12e257e-7887-4795-9221-4db8fd5856dd\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.682326 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-log-socket\") pod \"b12e257e-7887-4795-9221-4db8fd5856dd\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.682340 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-cni-bin\") pod \"b12e257e-7887-4795-9221-4db8fd5856dd\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.682353 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-run-ovn-kubernetes\") pod \"b12e257e-7887-4795-9221-4db8fd5856dd\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.682376 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-node-log\") pod \"b12e257e-7887-4795-9221-4db8fd5856dd\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.682403 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-run-ovn\") pod \"b12e257e-7887-4795-9221-4db8fd5856dd\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.682417 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b12e257e-7887-4795-9221-4db8fd5856dd-ovnkube-script-lib\") pod \"b12e257e-7887-4795-9221-4db8fd5856dd\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.682433 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jcdbt\" (UniqueName: \"kubernetes.io/projected/b12e257e-7887-4795-9221-4db8fd5856dd-kube-api-access-jcdbt\") pod \"b12e257e-7887-4795-9221-4db8fd5856dd\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.682457 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-etc-openvswitch\") pod \"b12e257e-7887-4795-9221-4db8fd5856dd\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.682487 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b12e257e-7887-4795-9221-4db8fd5856dd-env-overrides\") pod \"b12e257e-7887-4795-9221-4db8fd5856dd\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.682504 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-run-systemd\") pod \"b12e257e-7887-4795-9221-4db8fd5856dd\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.682524 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-var-lib-openvswitch\") pod \"b12e257e-7887-4795-9221-4db8fd5856dd\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.682538 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-run-openvswitch\") pod \"b12e257e-7887-4795-9221-4db8fd5856dd\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.682550 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"b12e257e-7887-4795-9221-4db8fd5856dd\" (UID: \"b12e257e-7887-4795-9221-4db8fd5856dd\") " Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.682756 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "b12e257e-7887-4795-9221-4db8fd5856dd" (UID: "b12e257e-7887-4795-9221-4db8fd5856dd"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.682785 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "b12e257e-7887-4795-9221-4db8fd5856dd" (UID: "b12e257e-7887-4795-9221-4db8fd5856dd"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.682804 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "b12e257e-7887-4795-9221-4db8fd5856dd" (UID: "b12e257e-7887-4795-9221-4db8fd5856dd"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.683152 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b12e257e-7887-4795-9221-4db8fd5856dd-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "b12e257e-7887-4795-9221-4db8fd5856dd" (UID: "b12e257e-7887-4795-9221-4db8fd5856dd"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.683185 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "b12e257e-7887-4795-9221-4db8fd5856dd" (UID: "b12e257e-7887-4795-9221-4db8fd5856dd"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.683414 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "b12e257e-7887-4795-9221-4db8fd5856dd" (UID: "b12e257e-7887-4795-9221-4db8fd5856dd"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.683463 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-slash" (OuterVolumeSpecName: "host-slash") pod "b12e257e-7887-4795-9221-4db8fd5856dd" (UID: "b12e257e-7887-4795-9221-4db8fd5856dd"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.683496 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-log-socket" (OuterVolumeSpecName: "log-socket") pod "b12e257e-7887-4795-9221-4db8fd5856dd" (UID: "b12e257e-7887-4795-9221-4db8fd5856dd"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.683515 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "b12e257e-7887-4795-9221-4db8fd5856dd" (UID: "b12e257e-7887-4795-9221-4db8fd5856dd"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.683531 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "b12e257e-7887-4795-9221-4db8fd5856dd" (UID: "b12e257e-7887-4795-9221-4db8fd5856dd"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.683549 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-node-log" (OuterVolumeSpecName: "node-log") pod "b12e257e-7887-4795-9221-4db8fd5856dd" (UID: "b12e257e-7887-4795-9221-4db8fd5856dd"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.683566 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "b12e257e-7887-4795-9221-4db8fd5856dd" (UID: "b12e257e-7887-4795-9221-4db8fd5856dd"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.683871 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b12e257e-7887-4795-9221-4db8fd5856dd-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "b12e257e-7887-4795-9221-4db8fd5856dd" (UID: "b12e257e-7887-4795-9221-4db8fd5856dd"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.683976 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "b12e257e-7887-4795-9221-4db8fd5856dd" (UID: "b12e257e-7887-4795-9221-4db8fd5856dd"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.684034 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "b12e257e-7887-4795-9221-4db8fd5856dd" (UID: "b12e257e-7887-4795-9221-4db8fd5856dd"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.684084 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b12e257e-7887-4795-9221-4db8fd5856dd-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "b12e257e-7887-4795-9221-4db8fd5856dd" (UID: "b12e257e-7887-4795-9221-4db8fd5856dd"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.684106 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "b12e257e-7887-4795-9221-4db8fd5856dd" (UID: "b12e257e-7887-4795-9221-4db8fd5856dd"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.688022 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b12e257e-7887-4795-9221-4db8fd5856dd-kube-api-access-jcdbt" (OuterVolumeSpecName: "kube-api-access-jcdbt") pod "b12e257e-7887-4795-9221-4db8fd5856dd" (UID: "b12e257e-7887-4795-9221-4db8fd5856dd"). InnerVolumeSpecName "kube-api-access-jcdbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.688986 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b12e257e-7887-4795-9221-4db8fd5856dd-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "b12e257e-7887-4795-9221-4db8fd5856dd" (UID: "b12e257e-7887-4795-9221-4db8fd5856dd"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.694732 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "b12e257e-7887-4795-9221-4db8fd5856dd" (UID: "b12e257e-7887-4795-9221-4db8fd5856dd"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784107 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-systemd-units\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784148 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-host-cni-netd\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784167 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0b38d706-f2ee-46e5-a84e-78da0f7b221c-ovnkube-script-lib\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784183 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0b38d706-f2ee-46e5-a84e-78da0f7b221c-ovn-node-metrics-cert\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784316 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0b38d706-f2ee-46e5-a84e-78da0f7b221c-env-overrides\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784337 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0b38d706-f2ee-46e5-a84e-78da0f7b221c-ovnkube-config\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784359 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-etc-openvswitch\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784400 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-host-cni-bin\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784423 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-run-systemd\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784447 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784466 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wxsv\" (UniqueName: \"kubernetes.io/projected/0b38d706-f2ee-46e5-a84e-78da0f7b221c-kube-api-access-5wxsv\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784492 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-run-openvswitch\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784506 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-host-run-ovn-kubernetes\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784546 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-host-kubelet\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784567 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-node-log\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784580 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-var-lib-openvswitch\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784603 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-host-slash\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784624 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-host-run-netns\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784637 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-run-ovn\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784672 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-log-socket\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784717 4559 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784728 4559 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-slash\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784736 4559 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-log-socket\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784744 4559 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784752 4559 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784759 4559 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-node-log\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784767 4559 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784847 4559 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b12e257e-7887-4795-9221-4db8fd5856dd-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784878 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jcdbt\" (UniqueName: \"kubernetes.io/projected/b12e257e-7887-4795-9221-4db8fd5856dd-kube-api-access-jcdbt\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784894 4559 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784903 4559 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b12e257e-7887-4795-9221-4db8fd5856dd-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784916 4559 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784926 4559 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784936 4559 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784946 4559 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784957 4559 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784967 4559 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784976 4559 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b12e257e-7887-4795-9221-4db8fd5856dd-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784986 4559 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b12e257e-7887-4795-9221-4db8fd5856dd-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.784996 4559 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b12e257e-7887-4795-9221-4db8fd5856dd-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.813834 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qn4h8_18091d80-4837-4894-8583-9a2bd30e10b9/kube-multus/2.log" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.814149 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qn4h8_18091d80-4837-4894-8583-9a2bd30e10b9/kube-multus/1.log" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.814187 4559 generic.go:334] "Generic (PLEG): container finished" podID="18091d80-4837-4894-8583-9a2bd30e10b9" containerID="43e22f4ac21e7304b67c5d85e3daca35952d0af501662a9f20bff79f4b065ff7" exitCode=2 Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.814233 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qn4h8" event={"ID":"18091d80-4837-4894-8583-9a2bd30e10b9","Type":"ContainerDied","Data":"43e22f4ac21e7304b67c5d85e3daca35952d0af501662a9f20bff79f4b065ff7"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.814264 4559 scope.go:117] "RemoveContainer" containerID="73ab4ccc28f84b0b246d2a4464c101f4f3113b2e910eea31f0effd6b89a0a71f" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.814621 4559 scope.go:117] "RemoveContainer" containerID="43e22f4ac21e7304b67c5d85e3daca35952d0af501662a9f20bff79f4b065ff7" Nov 23 06:53:39 crc kubenswrapper[4559]: E1123 06:53:39.814813 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-qn4h8_openshift-multus(18091d80-4837-4894-8583-9a2bd30e10b9)\"" pod="openshift-multus/multus-qn4h8" podUID="18091d80-4837-4894-8583-9a2bd30e10b9" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.816701 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jt56f_b12e257e-7887-4795-9221-4db8fd5856dd/ovnkube-controller/3.log" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.818769 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jt56f_b12e257e-7887-4795-9221-4db8fd5856dd/ovn-acl-logging/0.log" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819172 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jt56f_b12e257e-7887-4795-9221-4db8fd5856dd/ovn-controller/0.log" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819503 4559 generic.go:334] "Generic (PLEG): container finished" podID="b12e257e-7887-4795-9221-4db8fd5856dd" containerID="0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8" exitCode=0 Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819521 4559 generic.go:334] "Generic (PLEG): container finished" podID="b12e257e-7887-4795-9221-4db8fd5856dd" containerID="52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90" exitCode=0 Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819529 4559 generic.go:334] "Generic (PLEG): container finished" podID="b12e257e-7887-4795-9221-4db8fd5856dd" containerID="1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb" exitCode=0 Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819535 4559 generic.go:334] "Generic (PLEG): container finished" podID="b12e257e-7887-4795-9221-4db8fd5856dd" containerID="664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3" exitCode=0 Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819534 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" event={"ID":"b12e257e-7887-4795-9221-4db8fd5856dd","Type":"ContainerDied","Data":"0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819581 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" event={"ID":"b12e257e-7887-4795-9221-4db8fd5856dd","Type":"ContainerDied","Data":"52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819586 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819593 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" event={"ID":"b12e257e-7887-4795-9221-4db8fd5856dd","Type":"ContainerDied","Data":"1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819604 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" event={"ID":"b12e257e-7887-4795-9221-4db8fd5856dd","Type":"ContainerDied","Data":"664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819612 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" event={"ID":"b12e257e-7887-4795-9221-4db8fd5856dd","Type":"ContainerDied","Data":"36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819544 4559 generic.go:334] "Generic (PLEG): container finished" podID="b12e257e-7887-4795-9221-4db8fd5856dd" containerID="36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d" exitCode=0 Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819632 4559 generic.go:334] "Generic (PLEG): container finished" podID="b12e257e-7887-4795-9221-4db8fd5856dd" containerID="ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77" exitCode=0 Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819676 4559 generic.go:334] "Generic (PLEG): container finished" podID="b12e257e-7887-4795-9221-4db8fd5856dd" containerID="651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565" exitCode=143 Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819688 4559 generic.go:334] "Generic (PLEG): container finished" podID="b12e257e-7887-4795-9221-4db8fd5856dd" containerID="96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb" exitCode=143 Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819683 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" event={"ID":"b12e257e-7887-4795-9221-4db8fd5856dd","Type":"ContainerDied","Data":"ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819711 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819720 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819726 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819731 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819746 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819751 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819756 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819760 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819765 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819769 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819778 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" event={"ID":"b12e257e-7887-4795-9221-4db8fd5856dd","Type":"ContainerDied","Data":"651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819786 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819791 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819796 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819801 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819805 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819809 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819815 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819819 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819825 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819830 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819837 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" event={"ID":"b12e257e-7887-4795-9221-4db8fd5856dd","Type":"ContainerDied","Data":"96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819844 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819851 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819856 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819861 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819866 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819870 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819874 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819879 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819884 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819888 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819894 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jt56f" event={"ID":"b12e257e-7887-4795-9221-4db8fd5856dd","Type":"ContainerDied","Data":"f87f51d90c99a905784a62114224a7ff2e76c7d1929991399bda41703cc39d45"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819900 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819905 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819909 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819914 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819919 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819923 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819927 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819931 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819936 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.819940 4559 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d"} Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.833337 4559 scope.go:117] "RemoveContainer" containerID="0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.844456 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jt56f"] Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.846735 4559 scope.go:117] "RemoveContainer" containerID="28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.847444 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jt56f"] Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.878601 4559 scope.go:117] "RemoveContainer" containerID="52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.886106 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-host-cni-netd\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.886197 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-host-cni-netd\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.886247 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0b38d706-f2ee-46e5-a84e-78da0f7b221c-ovnkube-script-lib\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.886280 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0b38d706-f2ee-46e5-a84e-78da0f7b221c-ovn-node-metrics-cert\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.887164 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0b38d706-f2ee-46e5-a84e-78da0f7b221c-env-overrides\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.887191 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0b38d706-f2ee-46e5-a84e-78da0f7b221c-ovnkube-config\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.887211 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-etc-openvswitch\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.887256 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-host-cni-bin\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.887281 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-run-systemd\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.887079 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0b38d706-f2ee-46e5-a84e-78da0f7b221c-ovnkube-script-lib\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.887429 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-host-cni-bin\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.887611 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.887670 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-run-openvswitch\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.887673 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-run-systemd\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.887691 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wxsv\" (UniqueName: \"kubernetes.io/projected/0b38d706-f2ee-46e5-a84e-78da0f7b221c-kube-api-access-5wxsv\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.887710 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-host-run-ovn-kubernetes\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.887715 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.887744 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-run-openvswitch\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.887769 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-host-run-ovn-kubernetes\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.887826 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-host-kubelet\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.887887 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-node-log\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.887909 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-var-lib-openvswitch\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.887944 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-host-slash\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.887975 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-node-log\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.887980 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0b38d706-f2ee-46e5-a84e-78da0f7b221c-env-overrides\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.887994 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-host-run-netns\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.888001 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-host-kubelet\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.887980 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-host-run-netns\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.888020 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-var-lib-openvswitch\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.888024 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-host-slash\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.888030 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-log-socket\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.888044 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-log-socket\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.888069 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-etc-openvswitch\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.888088 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-run-ovn\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.888105 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-run-ovn\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.888139 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-systemd-units\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.888196 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0b38d706-f2ee-46e5-a84e-78da0f7b221c-systemd-units\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.888373 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0b38d706-f2ee-46e5-a84e-78da0f7b221c-ovnkube-config\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.889361 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0b38d706-f2ee-46e5-a84e-78da0f7b221c-ovn-node-metrics-cert\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.891773 4559 scope.go:117] "RemoveContainer" containerID="1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.901668 4559 scope.go:117] "RemoveContainer" containerID="664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.902503 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wxsv\" (UniqueName: \"kubernetes.io/projected/0b38d706-f2ee-46e5-a84e-78da0f7b221c-kube-api-access-5wxsv\") pod \"ovnkube-node-lv845\" (UID: \"0b38d706-f2ee-46e5-a84e-78da0f7b221c\") " pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.910793 4559 scope.go:117] "RemoveContainer" containerID="36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.920329 4559 scope.go:117] "RemoveContainer" containerID="ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.929284 4559 scope.go:117] "RemoveContainer" containerID="651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.930207 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.942161 4559 scope.go:117] "RemoveContainer" containerID="96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb" Nov 23 06:53:39 crc kubenswrapper[4559]: W1123 06:53:39.949737 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0b38d706_f2ee_46e5_a84e_78da0f7b221c.slice/crio-c8aeb1ea5614c9fe136c2758e18a583e3313be71ac85cef5bb68adb95b6a3b04 WatchSource:0}: Error finding container c8aeb1ea5614c9fe136c2758e18a583e3313be71ac85cef5bb68adb95b6a3b04: Status 404 returned error can't find the container with id c8aeb1ea5614c9fe136c2758e18a583e3313be71ac85cef5bb68adb95b6a3b04 Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.953250 4559 scope.go:117] "RemoveContainer" containerID="5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.966466 4559 scope.go:117] "RemoveContainer" containerID="0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8" Nov 23 06:53:39 crc kubenswrapper[4559]: E1123 06:53:39.966893 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8\": container with ID starting with 0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8 not found: ID does not exist" containerID="0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.966924 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8"} err="failed to get container status \"0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8\": rpc error: code = NotFound desc = could not find container \"0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8\": container with ID starting with 0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8 not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.966943 4559 scope.go:117] "RemoveContainer" containerID="28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da" Nov 23 06:53:39 crc kubenswrapper[4559]: E1123 06:53:39.967551 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da\": container with ID starting with 28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da not found: ID does not exist" containerID="28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.967584 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da"} err="failed to get container status \"28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da\": rpc error: code = NotFound desc = could not find container \"28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da\": container with ID starting with 28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.967605 4559 scope.go:117] "RemoveContainer" containerID="52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90" Nov 23 06:53:39 crc kubenswrapper[4559]: E1123 06:53:39.967863 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\": container with ID starting with 52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90 not found: ID does not exist" containerID="52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.967887 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90"} err="failed to get container status \"52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\": rpc error: code = NotFound desc = could not find container \"52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\": container with ID starting with 52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90 not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.967903 4559 scope.go:117] "RemoveContainer" containerID="1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb" Nov 23 06:53:39 crc kubenswrapper[4559]: E1123 06:53:39.968204 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\": container with ID starting with 1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb not found: ID does not exist" containerID="1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.968228 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb"} err="failed to get container status \"1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\": rpc error: code = NotFound desc = could not find container \"1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\": container with ID starting with 1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.968240 4559 scope.go:117] "RemoveContainer" containerID="664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3" Nov 23 06:53:39 crc kubenswrapper[4559]: E1123 06:53:39.968465 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\": container with ID starting with 664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3 not found: ID does not exist" containerID="664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.968490 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3"} err="failed to get container status \"664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\": rpc error: code = NotFound desc = could not find container \"664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\": container with ID starting with 664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3 not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.968501 4559 scope.go:117] "RemoveContainer" containerID="36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d" Nov 23 06:53:39 crc kubenswrapper[4559]: E1123 06:53:39.968723 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\": container with ID starting with 36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d not found: ID does not exist" containerID="36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.968741 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d"} err="failed to get container status \"36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\": rpc error: code = NotFound desc = could not find container \"36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\": container with ID starting with 36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.968752 4559 scope.go:117] "RemoveContainer" containerID="ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77" Nov 23 06:53:39 crc kubenswrapper[4559]: E1123 06:53:39.968969 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\": container with ID starting with ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77 not found: ID does not exist" containerID="ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.968989 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77"} err="failed to get container status \"ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\": rpc error: code = NotFound desc = could not find container \"ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\": container with ID starting with ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77 not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.969001 4559 scope.go:117] "RemoveContainer" containerID="651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565" Nov 23 06:53:39 crc kubenswrapper[4559]: E1123 06:53:39.969207 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\": container with ID starting with 651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565 not found: ID does not exist" containerID="651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.969225 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565"} err="failed to get container status \"651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\": rpc error: code = NotFound desc = could not find container \"651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\": container with ID starting with 651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565 not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.969236 4559 scope.go:117] "RemoveContainer" containerID="96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb" Nov 23 06:53:39 crc kubenswrapper[4559]: E1123 06:53:39.969494 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\": container with ID starting with 96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb not found: ID does not exist" containerID="96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.969512 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb"} err="failed to get container status \"96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\": rpc error: code = NotFound desc = could not find container \"96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\": container with ID starting with 96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.969524 4559 scope.go:117] "RemoveContainer" containerID="5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d" Nov 23 06:53:39 crc kubenswrapper[4559]: E1123 06:53:39.969807 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\": container with ID starting with 5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d not found: ID does not exist" containerID="5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.969828 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d"} err="failed to get container status \"5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\": rpc error: code = NotFound desc = could not find container \"5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\": container with ID starting with 5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.969846 4559 scope.go:117] "RemoveContainer" containerID="0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.970115 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8"} err="failed to get container status \"0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8\": rpc error: code = NotFound desc = could not find container \"0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8\": container with ID starting with 0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8 not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.970130 4559 scope.go:117] "RemoveContainer" containerID="28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.970315 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da"} err="failed to get container status \"28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da\": rpc error: code = NotFound desc = could not find container \"28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da\": container with ID starting with 28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.970333 4559 scope.go:117] "RemoveContainer" containerID="52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.970634 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90"} err="failed to get container status \"52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\": rpc error: code = NotFound desc = could not find container \"52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\": container with ID starting with 52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90 not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.970671 4559 scope.go:117] "RemoveContainer" containerID="1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.970905 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb"} err="failed to get container status \"1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\": rpc error: code = NotFound desc = could not find container \"1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\": container with ID starting with 1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.970922 4559 scope.go:117] "RemoveContainer" containerID="664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.971103 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3"} err="failed to get container status \"664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\": rpc error: code = NotFound desc = could not find container \"664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\": container with ID starting with 664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3 not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.971124 4559 scope.go:117] "RemoveContainer" containerID="36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.971316 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d"} err="failed to get container status \"36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\": rpc error: code = NotFound desc = could not find container \"36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\": container with ID starting with 36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.971333 4559 scope.go:117] "RemoveContainer" containerID="ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.971508 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77"} err="failed to get container status \"ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\": rpc error: code = NotFound desc = could not find container \"ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\": container with ID starting with ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77 not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.971525 4559 scope.go:117] "RemoveContainer" containerID="651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.971725 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565"} err="failed to get container status \"651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\": rpc error: code = NotFound desc = could not find container \"651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\": container with ID starting with 651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565 not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.971741 4559 scope.go:117] "RemoveContainer" containerID="96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.971905 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb"} err="failed to get container status \"96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\": rpc error: code = NotFound desc = could not find container \"96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\": container with ID starting with 96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.971926 4559 scope.go:117] "RemoveContainer" containerID="5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.972207 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d"} err="failed to get container status \"5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\": rpc error: code = NotFound desc = could not find container \"5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\": container with ID starting with 5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.972224 4559 scope.go:117] "RemoveContainer" containerID="0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.972445 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8"} err="failed to get container status \"0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8\": rpc error: code = NotFound desc = could not find container \"0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8\": container with ID starting with 0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8 not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.972461 4559 scope.go:117] "RemoveContainer" containerID="28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.972725 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da"} err="failed to get container status \"28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da\": rpc error: code = NotFound desc = could not find container \"28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da\": container with ID starting with 28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.972746 4559 scope.go:117] "RemoveContainer" containerID="52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.972945 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90"} err="failed to get container status \"52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\": rpc error: code = NotFound desc = could not find container \"52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\": container with ID starting with 52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90 not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.972963 4559 scope.go:117] "RemoveContainer" containerID="1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.973171 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb"} err="failed to get container status \"1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\": rpc error: code = NotFound desc = could not find container \"1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\": container with ID starting with 1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.973187 4559 scope.go:117] "RemoveContainer" containerID="664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.973361 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3"} err="failed to get container status \"664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\": rpc error: code = NotFound desc = could not find container \"664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\": container with ID starting with 664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3 not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.973409 4559 scope.go:117] "RemoveContainer" containerID="36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.973602 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d"} err="failed to get container status \"36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\": rpc error: code = NotFound desc = could not find container \"36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\": container with ID starting with 36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.973619 4559 scope.go:117] "RemoveContainer" containerID="ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.973819 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77"} err="failed to get container status \"ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\": rpc error: code = NotFound desc = could not find container \"ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\": container with ID starting with ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77 not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.973837 4559 scope.go:117] "RemoveContainer" containerID="651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.974163 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565"} err="failed to get container status \"651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\": rpc error: code = NotFound desc = could not find container \"651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\": container with ID starting with 651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565 not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.974188 4559 scope.go:117] "RemoveContainer" containerID="96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.974467 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb"} err="failed to get container status \"96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\": rpc error: code = NotFound desc = could not find container \"96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\": container with ID starting with 96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.974493 4559 scope.go:117] "RemoveContainer" containerID="5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.974800 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d"} err="failed to get container status \"5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\": rpc error: code = NotFound desc = could not find container \"5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\": container with ID starting with 5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.974820 4559 scope.go:117] "RemoveContainer" containerID="0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.975175 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8"} err="failed to get container status \"0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8\": rpc error: code = NotFound desc = could not find container \"0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8\": container with ID starting with 0f1631599969691f58c43a8cf158a8e2f933951d781159d175e4dcd31e2600a8 not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.975193 4559 scope.go:117] "RemoveContainer" containerID="28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.975450 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da"} err="failed to get container status \"28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da\": rpc error: code = NotFound desc = could not find container \"28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da\": container with ID starting with 28156e38c8f0a20826a92e173d4e6c556fbd5a4ec3293d953869562e4e2365da not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.975479 4559 scope.go:117] "RemoveContainer" containerID="52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.975729 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90"} err="failed to get container status \"52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\": rpc error: code = NotFound desc = could not find container \"52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90\": container with ID starting with 52c2d729f8d7f186f1edfbae693fd131176be374a5e9070d8d4022e6e2d56f90 not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.975748 4559 scope.go:117] "RemoveContainer" containerID="1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.976035 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb"} err="failed to get container status \"1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\": rpc error: code = NotFound desc = could not find container \"1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb\": container with ID starting with 1deac60beec3089c851acc2b6b3f73f9c12cd37536958aa59c444935b3828ebb not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.976057 4559 scope.go:117] "RemoveContainer" containerID="664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.976294 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3"} err="failed to get container status \"664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\": rpc error: code = NotFound desc = could not find container \"664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3\": container with ID starting with 664f9fe356fbba4683d5643a1285dec3c3f800b06507d63d56b5cf8bc6d95cc3 not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.976311 4559 scope.go:117] "RemoveContainer" containerID="36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.976534 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d"} err="failed to get container status \"36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\": rpc error: code = NotFound desc = could not find container \"36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d\": container with ID starting with 36ca3b3d25f493a5f541b2af1555c1e71295a48fd47b493aaebad929fb31706d not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.976550 4559 scope.go:117] "RemoveContainer" containerID="ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.976888 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77"} err="failed to get container status \"ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\": rpc error: code = NotFound desc = could not find container \"ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77\": container with ID starting with ff297036fba2d2f6266948fad903058dbcd7bd396c23c3ff07fd04aaea317c77 not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.976917 4559 scope.go:117] "RemoveContainer" containerID="651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.977093 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565"} err="failed to get container status \"651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\": rpc error: code = NotFound desc = could not find container \"651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565\": container with ID starting with 651eb54bf3eac606c18d148ca532229a6da46cb817e40d53c2d28d9b062bf565 not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.977111 4559 scope.go:117] "RemoveContainer" containerID="96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.977484 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb"} err="failed to get container status \"96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\": rpc error: code = NotFound desc = could not find container \"96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb\": container with ID starting with 96d6876f5de70a45d8ccfe9fbf0fd2c5209bb9e4deab63432124e9cfe4b04cfb not found: ID does not exist" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.977500 4559 scope.go:117] "RemoveContainer" containerID="5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d" Nov 23 06:53:39 crc kubenswrapper[4559]: I1123 06:53:39.977758 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d"} err="failed to get container status \"5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\": rpc error: code = NotFound desc = could not find container \"5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d\": container with ID starting with 5c672211323358dc2a0e794e73c7e5b44e6876e3bbf13922475e1e949561ed5d not found: ID does not exist" Nov 23 06:53:40 crc kubenswrapper[4559]: I1123 06:53:40.278525 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b12e257e-7887-4795-9221-4db8fd5856dd" path="/var/lib/kubelet/pods/b12e257e-7887-4795-9221-4db8fd5856dd/volumes" Nov 23 06:53:40 crc kubenswrapper[4559]: I1123 06:53:40.825375 4559 generic.go:334] "Generic (PLEG): container finished" podID="0b38d706-f2ee-46e5-a84e-78da0f7b221c" containerID="848f9899caf8d0c247bfa464a2bc8b9daba6da204a7e085c6de1779204e0e4a4" exitCode=0 Nov 23 06:53:40 crc kubenswrapper[4559]: I1123 06:53:40.825457 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lv845" event={"ID":"0b38d706-f2ee-46e5-a84e-78da0f7b221c","Type":"ContainerDied","Data":"848f9899caf8d0c247bfa464a2bc8b9daba6da204a7e085c6de1779204e0e4a4"} Nov 23 06:53:40 crc kubenswrapper[4559]: I1123 06:53:40.825516 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lv845" event={"ID":"0b38d706-f2ee-46e5-a84e-78da0f7b221c","Type":"ContainerStarted","Data":"c8aeb1ea5614c9fe136c2758e18a583e3313be71ac85cef5bb68adb95b6a3b04"} Nov 23 06:53:40 crc kubenswrapper[4559]: I1123 06:53:40.827822 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qn4h8_18091d80-4837-4894-8583-9a2bd30e10b9/kube-multus/2.log" Nov 23 06:53:41 crc kubenswrapper[4559]: I1123 06:53:41.835015 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lv845" event={"ID":"0b38d706-f2ee-46e5-a84e-78da0f7b221c","Type":"ContainerStarted","Data":"e761ea79abc85f75b1fd5537a70c6e7fbd04082ac2dd337c55c3d5ab3abab5e1"} Nov 23 06:53:41 crc kubenswrapper[4559]: I1123 06:53:41.835414 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lv845" event={"ID":"0b38d706-f2ee-46e5-a84e-78da0f7b221c","Type":"ContainerStarted","Data":"80bfacfc22f36e6868c97ade2daa2c7f9255911e873c328a28d1c474e8f586af"} Nov 23 06:53:41 crc kubenswrapper[4559]: I1123 06:53:41.835425 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lv845" event={"ID":"0b38d706-f2ee-46e5-a84e-78da0f7b221c","Type":"ContainerStarted","Data":"1d70ff0a685ace9a74aecc04a2d0ac5378d7425a13c53ae6d04b5c4b4ac01c52"} Nov 23 06:53:41 crc kubenswrapper[4559]: I1123 06:53:41.835433 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lv845" event={"ID":"0b38d706-f2ee-46e5-a84e-78da0f7b221c","Type":"ContainerStarted","Data":"c9fc7d6ca85fa8c65fcd5e8e1aeceb91869b26266787bc3b6dd9020974037cf1"} Nov 23 06:53:41 crc kubenswrapper[4559]: I1123 06:53:41.835440 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lv845" event={"ID":"0b38d706-f2ee-46e5-a84e-78da0f7b221c","Type":"ContainerStarted","Data":"89dbf4391355552f662586463331e5f04bb2aaf9efe85ab7afd48c3fc44de90a"} Nov 23 06:53:41 crc kubenswrapper[4559]: I1123 06:53:41.835448 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lv845" event={"ID":"0b38d706-f2ee-46e5-a84e-78da0f7b221c","Type":"ContainerStarted","Data":"e9ed763a860e209908e9fba1b36bbd13f34d9c756af845de6aae084d6a7ab2fe"} Nov 23 06:53:43 crc kubenswrapper[4559]: I1123 06:53:43.847752 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lv845" event={"ID":"0b38d706-f2ee-46e5-a84e-78da0f7b221c","Type":"ContainerStarted","Data":"a326503f4613e3be015b6cb800610231dadfae08c5f36f95f0aa1168c6000098"} Nov 23 06:53:45 crc kubenswrapper[4559]: I1123 06:53:45.860056 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lv845" event={"ID":"0b38d706-f2ee-46e5-a84e-78da0f7b221c","Type":"ContainerStarted","Data":"234931f2c91bf9cfd3c6b94d3607daa57cb5a77de7d39760a6f4a703a92cc54a"} Nov 23 06:53:45 crc kubenswrapper[4559]: I1123 06:53:45.860328 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:45 crc kubenswrapper[4559]: I1123 06:53:45.860358 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:45 crc kubenswrapper[4559]: I1123 06:53:45.888275 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:45 crc kubenswrapper[4559]: I1123 06:53:45.888682 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-lv845" podStartSLOduration=6.888668104 podStartE2EDuration="6.888668104s" podCreationTimestamp="2025-11-23 06:53:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:53:45.887480206 +0000 UTC m=+527.909465820" watchObservedRunningTime="2025-11-23 06:53:45.888668104 +0000 UTC m=+527.910653717" Nov 23 06:53:46 crc kubenswrapper[4559]: I1123 06:53:46.864742 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:46 crc kubenswrapper[4559]: I1123 06:53:46.886928 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:53:53 crc kubenswrapper[4559]: I1123 06:53:53.274841 4559 scope.go:117] "RemoveContainer" containerID="43e22f4ac21e7304b67c5d85e3daca35952d0af501662a9f20bff79f4b065ff7" Nov 23 06:53:53 crc kubenswrapper[4559]: E1123 06:53:53.275533 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-qn4h8_openshift-multus(18091d80-4837-4894-8583-9a2bd30e10b9)\"" pod="openshift-multus/multus-qn4h8" podUID="18091d80-4837-4894-8583-9a2bd30e10b9" Nov 23 06:53:56 crc kubenswrapper[4559]: I1123 06:53:56.167457 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 06:53:56 crc kubenswrapper[4559]: I1123 06:53:56.167875 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 06:53:58 crc kubenswrapper[4559]: I1123 06:53:58.356762 4559 scope.go:117] "RemoveContainer" containerID="cfd840e030d50c634bf5c50ec46181610c44cf82ed4ac54e427cc91d197fc0e6" Nov 23 06:53:58 crc kubenswrapper[4559]: I1123 06:53:58.371347 4559 scope.go:117] "RemoveContainer" containerID="3fbae69bb1331f980e3710ca233be9b9d5bf936ee95a62c9a0598863aed0654f" Nov 23 06:53:58 crc kubenswrapper[4559]: I1123 06:53:58.386776 4559 scope.go:117] "RemoveContainer" containerID="73b67ebd55b02d6deacf6718d5f6047c50a55e888094a012afe1ccac9b3fdd9e" Nov 23 06:53:58 crc kubenswrapper[4559]: I1123 06:53:58.396499 4559 scope.go:117] "RemoveContainer" containerID="583d1a00ef704e1ed48d88680d0fd97403d80d0a5354bf15dbb8f6eaba7d20e0" Nov 23 06:53:58 crc kubenswrapper[4559]: I1123 06:53:58.406783 4559 scope.go:117] "RemoveContainer" containerID="751894a83450d6c28894578e2c2c26826c5da6a0d6bce77bb13bc2ed45643811" Nov 23 06:53:58 crc kubenswrapper[4559]: I1123 06:53:58.421237 4559 scope.go:117] "RemoveContainer" containerID="7f65ec22cc44f19a9f306c55c9921d313d8ba7c45edfe2d63cb4f3b1dd40d5cd" Nov 23 06:54:04 crc kubenswrapper[4559]: I1123 06:54:04.275598 4559 scope.go:117] "RemoveContainer" containerID="43e22f4ac21e7304b67c5d85e3daca35952d0af501662a9f20bff79f4b065ff7" Nov 23 06:54:04 crc kubenswrapper[4559]: I1123 06:54:04.852832 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg"] Nov 23 06:54:04 crc kubenswrapper[4559]: I1123 06:54:04.853907 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg" Nov 23 06:54:04 crc kubenswrapper[4559]: I1123 06:54:04.855362 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 23 06:54:04 crc kubenswrapper[4559]: I1123 06:54:04.859720 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg"] Nov 23 06:54:04 crc kubenswrapper[4559]: I1123 06:54:04.951613 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/44c771c2-63aa-45f3-8551-f6b127f6121c-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg\" (UID: \"44c771c2-63aa-45f3-8551-f6b127f6121c\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg" Nov 23 06:54:04 crc kubenswrapper[4559]: I1123 06:54:04.951811 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x46w2\" (UniqueName: \"kubernetes.io/projected/44c771c2-63aa-45f3-8551-f6b127f6121c-kube-api-access-x46w2\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg\" (UID: \"44c771c2-63aa-45f3-8551-f6b127f6121c\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg" Nov 23 06:54:04 crc kubenswrapper[4559]: I1123 06:54:04.951850 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/44c771c2-63aa-45f3-8551-f6b127f6121c-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg\" (UID: \"44c771c2-63aa-45f3-8551-f6b127f6121c\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg" Nov 23 06:54:04 crc kubenswrapper[4559]: I1123 06:54:04.963064 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qn4h8_18091d80-4837-4894-8583-9a2bd30e10b9/kube-multus/2.log" Nov 23 06:54:04 crc kubenswrapper[4559]: I1123 06:54:04.963169 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qn4h8" event={"ID":"18091d80-4837-4894-8583-9a2bd30e10b9","Type":"ContainerStarted","Data":"c6f83e1e2890b7df9cf9241fddb52bd0d4d1c944ec1ec8c59d02631d82b6c80a"} Nov 23 06:54:05 crc kubenswrapper[4559]: I1123 06:54:05.053706 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x46w2\" (UniqueName: \"kubernetes.io/projected/44c771c2-63aa-45f3-8551-f6b127f6121c-kube-api-access-x46w2\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg\" (UID: \"44c771c2-63aa-45f3-8551-f6b127f6121c\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg" Nov 23 06:54:05 crc kubenswrapper[4559]: I1123 06:54:05.054204 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/44c771c2-63aa-45f3-8551-f6b127f6121c-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg\" (UID: \"44c771c2-63aa-45f3-8551-f6b127f6121c\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg" Nov 23 06:54:05 crc kubenswrapper[4559]: I1123 06:54:05.054268 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/44c771c2-63aa-45f3-8551-f6b127f6121c-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg\" (UID: \"44c771c2-63aa-45f3-8551-f6b127f6121c\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg" Nov 23 06:54:05 crc kubenswrapper[4559]: I1123 06:54:05.054726 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/44c771c2-63aa-45f3-8551-f6b127f6121c-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg\" (UID: \"44c771c2-63aa-45f3-8551-f6b127f6121c\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg" Nov 23 06:54:05 crc kubenswrapper[4559]: I1123 06:54:05.054780 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/44c771c2-63aa-45f3-8551-f6b127f6121c-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg\" (UID: \"44c771c2-63aa-45f3-8551-f6b127f6121c\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg" Nov 23 06:54:05 crc kubenswrapper[4559]: I1123 06:54:05.069898 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x46w2\" (UniqueName: \"kubernetes.io/projected/44c771c2-63aa-45f3-8551-f6b127f6121c-kube-api-access-x46w2\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg\" (UID: \"44c771c2-63aa-45f3-8551-f6b127f6121c\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg" Nov 23 06:54:05 crc kubenswrapper[4559]: I1123 06:54:05.168812 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg" Nov 23 06:54:05 crc kubenswrapper[4559]: E1123 06:54:05.193194 4559 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg_openshift-marketplace_44c771c2-63aa-45f3-8551-f6b127f6121c_0(a620970492b6e4486e591e3362372840056737201c4eed62585c854b9aed36b3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 23 06:54:05 crc kubenswrapper[4559]: E1123 06:54:05.193279 4559 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg_openshift-marketplace_44c771c2-63aa-45f3-8551-f6b127f6121c_0(a620970492b6e4486e591e3362372840056737201c4eed62585c854b9aed36b3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg" Nov 23 06:54:05 crc kubenswrapper[4559]: E1123 06:54:05.193308 4559 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg_openshift-marketplace_44c771c2-63aa-45f3-8551-f6b127f6121c_0(a620970492b6e4486e591e3362372840056737201c4eed62585c854b9aed36b3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg" Nov 23 06:54:05 crc kubenswrapper[4559]: E1123 06:54:05.193378 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg_openshift-marketplace(44c771c2-63aa-45f3-8551-f6b127f6121c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg_openshift-marketplace(44c771c2-63aa-45f3-8551-f6b127f6121c)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg_openshift-marketplace_44c771c2-63aa-45f3-8551-f6b127f6121c_0(a620970492b6e4486e591e3362372840056737201c4eed62585c854b9aed36b3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg" podUID="44c771c2-63aa-45f3-8551-f6b127f6121c" Nov 23 06:54:05 crc kubenswrapper[4559]: I1123 06:54:05.967240 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg" Nov 23 06:54:05 crc kubenswrapper[4559]: I1123 06:54:05.967547 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg" Nov 23 06:54:06 crc kubenswrapper[4559]: I1123 06:54:06.110621 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg"] Nov 23 06:54:06 crc kubenswrapper[4559]: I1123 06:54:06.973108 4559 generic.go:334] "Generic (PLEG): container finished" podID="44c771c2-63aa-45f3-8551-f6b127f6121c" containerID="aa35bfe4f487054ebef75beeddc594284bc0352c9a88658067db5e83fee947f1" exitCode=0 Nov 23 06:54:06 crc kubenswrapper[4559]: I1123 06:54:06.973151 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg" event={"ID":"44c771c2-63aa-45f3-8551-f6b127f6121c","Type":"ContainerDied","Data":"aa35bfe4f487054ebef75beeddc594284bc0352c9a88658067db5e83fee947f1"} Nov 23 06:54:06 crc kubenswrapper[4559]: I1123 06:54:06.973176 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg" event={"ID":"44c771c2-63aa-45f3-8551-f6b127f6121c","Type":"ContainerStarted","Data":"6ca972425bf442bcf2adb3029f1d338130409e494f33311c8d1bc71c81531583"} Nov 23 06:54:08 crc kubenswrapper[4559]: I1123 06:54:08.986833 4559 generic.go:334] "Generic (PLEG): container finished" podID="44c771c2-63aa-45f3-8551-f6b127f6121c" containerID="c23e95d47a2b0234e53911c92b4d95f305876ac7a312267b8f1cca3ef51a4f03" exitCode=0 Nov 23 06:54:08 crc kubenswrapper[4559]: I1123 06:54:08.986923 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg" event={"ID":"44c771c2-63aa-45f3-8551-f6b127f6121c","Type":"ContainerDied","Data":"c23e95d47a2b0234e53911c92b4d95f305876ac7a312267b8f1cca3ef51a4f03"} Nov 23 06:54:09 crc kubenswrapper[4559]: I1123 06:54:09.949927 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lv845" Nov 23 06:54:09 crc kubenswrapper[4559]: I1123 06:54:09.994538 4559 generic.go:334] "Generic (PLEG): container finished" podID="44c771c2-63aa-45f3-8551-f6b127f6121c" containerID="498665aa1ec5c2998d4867f419f4a00c52885d7dddfff281316e225dd5f82f39" exitCode=0 Nov 23 06:54:09 crc kubenswrapper[4559]: I1123 06:54:09.994604 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg" event={"ID":"44c771c2-63aa-45f3-8551-f6b127f6121c","Type":"ContainerDied","Data":"498665aa1ec5c2998d4867f419f4a00c52885d7dddfff281316e225dd5f82f39"} Nov 23 06:54:11 crc kubenswrapper[4559]: I1123 06:54:11.188321 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg" Nov 23 06:54:11 crc kubenswrapper[4559]: I1123 06:54:11.229132 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x46w2\" (UniqueName: \"kubernetes.io/projected/44c771c2-63aa-45f3-8551-f6b127f6121c-kube-api-access-x46w2\") pod \"44c771c2-63aa-45f3-8551-f6b127f6121c\" (UID: \"44c771c2-63aa-45f3-8551-f6b127f6121c\") " Nov 23 06:54:11 crc kubenswrapper[4559]: I1123 06:54:11.229299 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/44c771c2-63aa-45f3-8551-f6b127f6121c-bundle\") pod \"44c771c2-63aa-45f3-8551-f6b127f6121c\" (UID: \"44c771c2-63aa-45f3-8551-f6b127f6121c\") " Nov 23 06:54:11 crc kubenswrapper[4559]: I1123 06:54:11.229350 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/44c771c2-63aa-45f3-8551-f6b127f6121c-util\") pod \"44c771c2-63aa-45f3-8551-f6b127f6121c\" (UID: \"44c771c2-63aa-45f3-8551-f6b127f6121c\") " Nov 23 06:54:11 crc kubenswrapper[4559]: I1123 06:54:11.229817 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44c771c2-63aa-45f3-8551-f6b127f6121c-bundle" (OuterVolumeSpecName: "bundle") pod "44c771c2-63aa-45f3-8551-f6b127f6121c" (UID: "44c771c2-63aa-45f3-8551-f6b127f6121c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:54:11 crc kubenswrapper[4559]: I1123 06:54:11.234469 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44c771c2-63aa-45f3-8551-f6b127f6121c-kube-api-access-x46w2" (OuterVolumeSpecName: "kube-api-access-x46w2") pod "44c771c2-63aa-45f3-8551-f6b127f6121c" (UID: "44c771c2-63aa-45f3-8551-f6b127f6121c"). InnerVolumeSpecName "kube-api-access-x46w2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:54:11 crc kubenswrapper[4559]: I1123 06:54:11.239546 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44c771c2-63aa-45f3-8551-f6b127f6121c-util" (OuterVolumeSpecName: "util") pod "44c771c2-63aa-45f3-8551-f6b127f6121c" (UID: "44c771c2-63aa-45f3-8551-f6b127f6121c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:54:11 crc kubenswrapper[4559]: I1123 06:54:11.331384 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x46w2\" (UniqueName: \"kubernetes.io/projected/44c771c2-63aa-45f3-8551-f6b127f6121c-kube-api-access-x46w2\") on node \"crc\" DevicePath \"\"" Nov 23 06:54:11 crc kubenswrapper[4559]: I1123 06:54:11.331413 4559 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/44c771c2-63aa-45f3-8551-f6b127f6121c-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:54:11 crc kubenswrapper[4559]: I1123 06:54:11.331426 4559 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/44c771c2-63aa-45f3-8551-f6b127f6121c-util\") on node \"crc\" DevicePath \"\"" Nov 23 06:54:12 crc kubenswrapper[4559]: I1123 06:54:12.006591 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg" event={"ID":"44c771c2-63aa-45f3-8551-f6b127f6121c","Type":"ContainerDied","Data":"6ca972425bf442bcf2adb3029f1d338130409e494f33311c8d1bc71c81531583"} Nov 23 06:54:12 crc kubenswrapper[4559]: I1123 06:54:12.006631 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ca972425bf442bcf2adb3029f1d338130409e494f33311c8d1bc71c81531583" Nov 23 06:54:12 crc kubenswrapper[4559]: I1123 06:54:12.006637 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg" Nov 23 06:54:13 crc kubenswrapper[4559]: I1123 06:54:13.730761 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-pfk9b"] Nov 23 06:54:13 crc kubenswrapper[4559]: E1123 06:54:13.731290 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44c771c2-63aa-45f3-8551-f6b127f6121c" containerName="util" Nov 23 06:54:13 crc kubenswrapper[4559]: I1123 06:54:13.731306 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="44c771c2-63aa-45f3-8551-f6b127f6121c" containerName="util" Nov 23 06:54:13 crc kubenswrapper[4559]: E1123 06:54:13.731318 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44c771c2-63aa-45f3-8551-f6b127f6121c" containerName="extract" Nov 23 06:54:13 crc kubenswrapper[4559]: I1123 06:54:13.731324 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="44c771c2-63aa-45f3-8551-f6b127f6121c" containerName="extract" Nov 23 06:54:13 crc kubenswrapper[4559]: E1123 06:54:13.731332 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44c771c2-63aa-45f3-8551-f6b127f6121c" containerName="pull" Nov 23 06:54:13 crc kubenswrapper[4559]: I1123 06:54:13.731339 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="44c771c2-63aa-45f3-8551-f6b127f6121c" containerName="pull" Nov 23 06:54:13 crc kubenswrapper[4559]: I1123 06:54:13.731457 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="44c771c2-63aa-45f3-8551-f6b127f6121c" containerName="extract" Nov 23 06:54:13 crc kubenswrapper[4559]: I1123 06:54:13.731940 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-pfk9b" Nov 23 06:54:13 crc kubenswrapper[4559]: I1123 06:54:13.734374 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 23 06:54:13 crc kubenswrapper[4559]: I1123 06:54:13.735058 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 23 06:54:13 crc kubenswrapper[4559]: I1123 06:54:13.737446 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-b66mv" Nov 23 06:54:13 crc kubenswrapper[4559]: I1123 06:54:13.745267 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-pfk9b"] Nov 23 06:54:13 crc kubenswrapper[4559]: I1123 06:54:13.758151 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vkpj\" (UniqueName: \"kubernetes.io/projected/c7cb504c-5b1d-4d00-b5e9-6bcc6968ee84-kube-api-access-7vkpj\") pod \"nmstate-operator-557fdffb88-pfk9b\" (UID: \"c7cb504c-5b1d-4d00-b5e9-6bcc6968ee84\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-pfk9b" Nov 23 06:54:13 crc kubenswrapper[4559]: I1123 06:54:13.859278 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vkpj\" (UniqueName: \"kubernetes.io/projected/c7cb504c-5b1d-4d00-b5e9-6bcc6968ee84-kube-api-access-7vkpj\") pod \"nmstate-operator-557fdffb88-pfk9b\" (UID: \"c7cb504c-5b1d-4d00-b5e9-6bcc6968ee84\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-pfk9b" Nov 23 06:54:13 crc kubenswrapper[4559]: I1123 06:54:13.874377 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vkpj\" (UniqueName: \"kubernetes.io/projected/c7cb504c-5b1d-4d00-b5e9-6bcc6968ee84-kube-api-access-7vkpj\") pod \"nmstate-operator-557fdffb88-pfk9b\" (UID: \"c7cb504c-5b1d-4d00-b5e9-6bcc6968ee84\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-pfk9b" Nov 23 06:54:14 crc kubenswrapper[4559]: I1123 06:54:14.044270 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-pfk9b" Nov 23 06:54:14 crc kubenswrapper[4559]: I1123 06:54:14.429783 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-pfk9b"] Nov 23 06:54:15 crc kubenswrapper[4559]: I1123 06:54:15.023745 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-pfk9b" event={"ID":"c7cb504c-5b1d-4d00-b5e9-6bcc6968ee84","Type":"ContainerStarted","Data":"ccaf7d473beb7af3fe818e7d8ca3d9d530033287b899f907d950cb4c90ac8569"} Nov 23 06:54:17 crc kubenswrapper[4559]: I1123 06:54:17.040794 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-pfk9b" event={"ID":"c7cb504c-5b1d-4d00-b5e9-6bcc6968ee84","Type":"ContainerStarted","Data":"479764b91d398f468bd8b5dd60abe18d075213d93eddac16bdfa106719bb6d53"} Nov 23 06:54:17 crc kubenswrapper[4559]: I1123 06:54:17.054915 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-557fdffb88-pfk9b" podStartSLOduration=2.334971171 podStartE2EDuration="4.054870207s" podCreationTimestamp="2025-11-23 06:54:13 +0000 UTC" firstStartedPulling="2025-11-23 06:54:14.444521251 +0000 UTC m=+556.466506866" lastFinishedPulling="2025-11-23 06:54:16.164420287 +0000 UTC m=+558.186405902" observedRunningTime="2025-11-23 06:54:17.053944724 +0000 UTC m=+559.075930338" watchObservedRunningTime="2025-11-23 06:54:17.054870207 +0000 UTC m=+559.076855821" Nov 23 06:54:17 crc kubenswrapper[4559]: I1123 06:54:17.832975 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-rpd7c"] Nov 23 06:54:17 crc kubenswrapper[4559]: I1123 06:54:17.834154 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-rpd7c" Nov 23 06:54:17 crc kubenswrapper[4559]: I1123 06:54:17.836653 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-6sdh6" Nov 23 06:54:17 crc kubenswrapper[4559]: I1123 06:54:17.839557 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-hv2js"] Nov 23 06:54:17 crc kubenswrapper[4559]: I1123 06:54:17.840332 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-hv2js" Nov 23 06:54:17 crc kubenswrapper[4559]: W1123 06:54:17.842128 4559 reflector.go:561] object-"openshift-nmstate"/"openshift-nmstate-webhook": failed to list *v1.Secret: secrets "openshift-nmstate-webhook" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-nmstate": no relationship found between node 'crc' and this object Nov 23 06:54:17 crc kubenswrapper[4559]: E1123 06:54:17.842162 4559 reflector.go:158] "Unhandled Error" err="object-\"openshift-nmstate\"/\"openshift-nmstate-webhook\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openshift-nmstate-webhook\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-nmstate\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 23 06:54:17 crc kubenswrapper[4559]: I1123 06:54:17.850298 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-rpd7c"] Nov 23 06:54:17 crc kubenswrapper[4559]: I1123 06:54:17.856068 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-fglqj"] Nov 23 06:54:17 crc kubenswrapper[4559]: I1123 06:54:17.856802 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-fglqj" Nov 23 06:54:17 crc kubenswrapper[4559]: I1123 06:54:17.858984 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-hv2js"] Nov 23 06:54:17 crc kubenswrapper[4559]: I1123 06:54:17.910615 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/092f86a7-8e24-4a70-9af6-a4265c3688e5-nmstate-lock\") pod \"nmstate-handler-fglqj\" (UID: \"092f86a7-8e24-4a70-9af6-a4265c3688e5\") " pod="openshift-nmstate/nmstate-handler-fglqj" Nov 23 06:54:17 crc kubenswrapper[4559]: I1123 06:54:17.910712 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsz5x\" (UniqueName: \"kubernetes.io/projected/092f86a7-8e24-4a70-9af6-a4265c3688e5-kube-api-access-nsz5x\") pod \"nmstate-handler-fglqj\" (UID: \"092f86a7-8e24-4a70-9af6-a4265c3688e5\") " pod="openshift-nmstate/nmstate-handler-fglqj" Nov 23 06:54:17 crc kubenswrapper[4559]: I1123 06:54:17.910780 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r52f9\" (UniqueName: \"kubernetes.io/projected/6ff08b99-2dc5-4b9f-bef5-b0d34e76cb49-kube-api-access-r52f9\") pod \"nmstate-webhook-6b89b748d8-hv2js\" (UID: \"6ff08b99-2dc5-4b9f-bef5-b0d34e76cb49\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-hv2js" Nov 23 06:54:17 crc kubenswrapper[4559]: I1123 06:54:17.910806 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/6ff08b99-2dc5-4b9f-bef5-b0d34e76cb49-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-hv2js\" (UID: \"6ff08b99-2dc5-4b9f-bef5-b0d34e76cb49\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-hv2js" Nov 23 06:54:17 crc kubenswrapper[4559]: I1123 06:54:17.910843 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/092f86a7-8e24-4a70-9af6-a4265c3688e5-ovs-socket\") pod \"nmstate-handler-fglqj\" (UID: \"092f86a7-8e24-4a70-9af6-a4265c3688e5\") " pod="openshift-nmstate/nmstate-handler-fglqj" Nov 23 06:54:17 crc kubenswrapper[4559]: I1123 06:54:17.910867 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/092f86a7-8e24-4a70-9af6-a4265c3688e5-dbus-socket\") pod \"nmstate-handler-fglqj\" (UID: \"092f86a7-8e24-4a70-9af6-a4265c3688e5\") " pod="openshift-nmstate/nmstate-handler-fglqj" Nov 23 06:54:17 crc kubenswrapper[4559]: I1123 06:54:17.910910 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrj5v\" (UniqueName: \"kubernetes.io/projected/3755e847-69fa-47e4-93f9-fe15df377011-kube-api-access-xrj5v\") pod \"nmstate-metrics-5dcf9c57c5-rpd7c\" (UID: \"3755e847-69fa-47e4-93f9-fe15df377011\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-rpd7c" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.011705 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/6ff08b99-2dc5-4b9f-bef5-b0d34e76cb49-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-hv2js\" (UID: \"6ff08b99-2dc5-4b9f-bef5-b0d34e76cb49\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-hv2js" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.011746 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/092f86a7-8e24-4a70-9af6-a4265c3688e5-ovs-socket\") pod \"nmstate-handler-fglqj\" (UID: \"092f86a7-8e24-4a70-9af6-a4265c3688e5\") " pod="openshift-nmstate/nmstate-handler-fglqj" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.011779 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/092f86a7-8e24-4a70-9af6-a4265c3688e5-dbus-socket\") pod \"nmstate-handler-fglqj\" (UID: \"092f86a7-8e24-4a70-9af6-a4265c3688e5\") " pod="openshift-nmstate/nmstate-handler-fglqj" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.011812 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrj5v\" (UniqueName: \"kubernetes.io/projected/3755e847-69fa-47e4-93f9-fe15df377011-kube-api-access-xrj5v\") pod \"nmstate-metrics-5dcf9c57c5-rpd7c\" (UID: \"3755e847-69fa-47e4-93f9-fe15df377011\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-rpd7c" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.011831 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/092f86a7-8e24-4a70-9af6-a4265c3688e5-nmstate-lock\") pod \"nmstate-handler-fglqj\" (UID: \"092f86a7-8e24-4a70-9af6-a4265c3688e5\") " pod="openshift-nmstate/nmstate-handler-fglqj" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.011850 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsz5x\" (UniqueName: \"kubernetes.io/projected/092f86a7-8e24-4a70-9af6-a4265c3688e5-kube-api-access-nsz5x\") pod \"nmstate-handler-fglqj\" (UID: \"092f86a7-8e24-4a70-9af6-a4265c3688e5\") " pod="openshift-nmstate/nmstate-handler-fglqj" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.011865 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/092f86a7-8e24-4a70-9af6-a4265c3688e5-ovs-socket\") pod \"nmstate-handler-fglqj\" (UID: \"092f86a7-8e24-4a70-9af6-a4265c3688e5\") " pod="openshift-nmstate/nmstate-handler-fglqj" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.011887 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r52f9\" (UniqueName: \"kubernetes.io/projected/6ff08b99-2dc5-4b9f-bef5-b0d34e76cb49-kube-api-access-r52f9\") pod \"nmstate-webhook-6b89b748d8-hv2js\" (UID: \"6ff08b99-2dc5-4b9f-bef5-b0d34e76cb49\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-hv2js" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.011948 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/092f86a7-8e24-4a70-9af6-a4265c3688e5-nmstate-lock\") pod \"nmstate-handler-fglqj\" (UID: \"092f86a7-8e24-4a70-9af6-a4265c3688e5\") " pod="openshift-nmstate/nmstate-handler-fglqj" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.012142 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/092f86a7-8e24-4a70-9af6-a4265c3688e5-dbus-socket\") pod \"nmstate-handler-fglqj\" (UID: \"092f86a7-8e24-4a70-9af6-a4265c3688e5\") " pod="openshift-nmstate/nmstate-handler-fglqj" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.019258 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6qvx9"] Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.019957 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6qvx9" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.023681 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-t5jr5" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.023681 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.023711 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.034685 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrj5v\" (UniqueName: \"kubernetes.io/projected/3755e847-69fa-47e4-93f9-fe15df377011-kube-api-access-xrj5v\") pod \"nmstate-metrics-5dcf9c57c5-rpd7c\" (UID: \"3755e847-69fa-47e4-93f9-fe15df377011\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-rpd7c" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.034734 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r52f9\" (UniqueName: \"kubernetes.io/projected/6ff08b99-2dc5-4b9f-bef5-b0d34e76cb49-kube-api-access-r52f9\") pod \"nmstate-webhook-6b89b748d8-hv2js\" (UID: \"6ff08b99-2dc5-4b9f-bef5-b0d34e76cb49\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-hv2js" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.036963 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6qvx9"] Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.037476 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsz5x\" (UniqueName: \"kubernetes.io/projected/092f86a7-8e24-4a70-9af6-a4265c3688e5-kube-api-access-nsz5x\") pod \"nmstate-handler-fglqj\" (UID: \"092f86a7-8e24-4a70-9af6-a4265c3688e5\") " pod="openshift-nmstate/nmstate-handler-fglqj" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.148442 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-rpd7c" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.172837 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-fglqj" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.186061 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6765c6547-hbtvp"] Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.186736 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6765c6547-hbtvp" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.213397 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6765c6547-hbtvp"] Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.214037 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b0662ab8-3035-4b7b-b615-15cc0b963af4-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-6qvx9\" (UID: \"b0662ab8-3035-4b7b-b615-15cc0b963af4\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6qvx9" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.214155 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b0662ab8-3035-4b7b-b615-15cc0b963af4-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-6qvx9\" (UID: \"b0662ab8-3035-4b7b-b615-15cc0b963af4\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6qvx9" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.214188 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcnwv\" (UniqueName: \"kubernetes.io/projected/b0662ab8-3035-4b7b-b615-15cc0b963af4-kube-api-access-wcnwv\") pod \"nmstate-console-plugin-5874bd7bc5-6qvx9\" (UID: \"b0662ab8-3035-4b7b-b615-15cc0b963af4\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6qvx9" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.315541 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88ed011c-aba3-49b4-9179-9bfc6fdd8ff8-trusted-ca-bundle\") pod \"console-6765c6547-hbtvp\" (UID: \"88ed011c-aba3-49b4-9179-9bfc6fdd8ff8\") " pod="openshift-console/console-6765c6547-hbtvp" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.315606 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b0662ab8-3035-4b7b-b615-15cc0b963af4-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-6qvx9\" (UID: \"b0662ab8-3035-4b7b-b615-15cc0b963af4\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6qvx9" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.315708 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/88ed011c-aba3-49b4-9179-9bfc6fdd8ff8-console-config\") pod \"console-6765c6547-hbtvp\" (UID: \"88ed011c-aba3-49b4-9179-9bfc6fdd8ff8\") " pod="openshift-console/console-6765c6547-hbtvp" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.316041 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/88ed011c-aba3-49b4-9179-9bfc6fdd8ff8-console-serving-cert\") pod \"console-6765c6547-hbtvp\" (UID: \"88ed011c-aba3-49b4-9179-9bfc6fdd8ff8\") " pod="openshift-console/console-6765c6547-hbtvp" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.316091 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/88ed011c-aba3-49b4-9179-9bfc6fdd8ff8-oauth-serving-cert\") pod \"console-6765c6547-hbtvp\" (UID: \"88ed011c-aba3-49b4-9179-9bfc6fdd8ff8\") " pod="openshift-console/console-6765c6547-hbtvp" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.316121 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcnwv\" (UniqueName: \"kubernetes.io/projected/b0662ab8-3035-4b7b-b615-15cc0b963af4-kube-api-access-wcnwv\") pod \"nmstate-console-plugin-5874bd7bc5-6qvx9\" (UID: \"b0662ab8-3035-4b7b-b615-15cc0b963af4\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6qvx9" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.316196 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/88ed011c-aba3-49b4-9179-9bfc6fdd8ff8-service-ca\") pod \"console-6765c6547-hbtvp\" (UID: \"88ed011c-aba3-49b4-9179-9bfc6fdd8ff8\") " pod="openshift-console/console-6765c6547-hbtvp" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.316266 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b0662ab8-3035-4b7b-b615-15cc0b963af4-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-6qvx9\" (UID: \"b0662ab8-3035-4b7b-b615-15cc0b963af4\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6qvx9" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.316296 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgndv\" (UniqueName: \"kubernetes.io/projected/88ed011c-aba3-49b4-9179-9bfc6fdd8ff8-kube-api-access-qgndv\") pod \"console-6765c6547-hbtvp\" (UID: \"88ed011c-aba3-49b4-9179-9bfc6fdd8ff8\") " pod="openshift-console/console-6765c6547-hbtvp" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.316328 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/88ed011c-aba3-49b4-9179-9bfc6fdd8ff8-console-oauth-config\") pod \"console-6765c6547-hbtvp\" (UID: \"88ed011c-aba3-49b4-9179-9bfc6fdd8ff8\") " pod="openshift-console/console-6765c6547-hbtvp" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.316335 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b0662ab8-3035-4b7b-b615-15cc0b963af4-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-6qvx9\" (UID: \"b0662ab8-3035-4b7b-b615-15cc0b963af4\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6qvx9" Nov 23 06:54:18 crc kubenswrapper[4559]: E1123 06:54:18.316403 4559 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Nov 23 06:54:18 crc kubenswrapper[4559]: E1123 06:54:18.316452 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b0662ab8-3035-4b7b-b615-15cc0b963af4-plugin-serving-cert podName:b0662ab8-3035-4b7b-b615-15cc0b963af4 nodeName:}" failed. No retries permitted until 2025-11-23 06:54:18.816437676 +0000 UTC m=+560.838423289 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/b0662ab8-3035-4b7b-b615-15cc0b963af4-plugin-serving-cert") pod "nmstate-console-plugin-5874bd7bc5-6qvx9" (UID: "b0662ab8-3035-4b7b-b615-15cc0b963af4") : secret "plugin-serving-cert" not found Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.333476 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcnwv\" (UniqueName: \"kubernetes.io/projected/b0662ab8-3035-4b7b-b615-15cc0b963af4-kube-api-access-wcnwv\") pod \"nmstate-console-plugin-5874bd7bc5-6qvx9\" (UID: \"b0662ab8-3035-4b7b-b615-15cc0b963af4\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6qvx9" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.386314 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-rpd7c"] Nov 23 06:54:18 crc kubenswrapper[4559]: W1123 06:54:18.390609 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3755e847_69fa_47e4_93f9_fe15df377011.slice/crio-8ecf552b7491d1fa3e0d6756b7fcca89c6b50b412f6bad392be072baacad8454 WatchSource:0}: Error finding container 8ecf552b7491d1fa3e0d6756b7fcca89c6b50b412f6bad392be072baacad8454: Status 404 returned error can't find the container with id 8ecf552b7491d1fa3e0d6756b7fcca89c6b50b412f6bad392be072baacad8454 Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.417710 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/88ed011c-aba3-49b4-9179-9bfc6fdd8ff8-service-ca\") pod \"console-6765c6547-hbtvp\" (UID: \"88ed011c-aba3-49b4-9179-9bfc6fdd8ff8\") " pod="openshift-console/console-6765c6547-hbtvp" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.417823 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgndv\" (UniqueName: \"kubernetes.io/projected/88ed011c-aba3-49b4-9179-9bfc6fdd8ff8-kube-api-access-qgndv\") pod \"console-6765c6547-hbtvp\" (UID: \"88ed011c-aba3-49b4-9179-9bfc6fdd8ff8\") " pod="openshift-console/console-6765c6547-hbtvp" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.417847 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/88ed011c-aba3-49b4-9179-9bfc6fdd8ff8-console-oauth-config\") pod \"console-6765c6547-hbtvp\" (UID: \"88ed011c-aba3-49b4-9179-9bfc6fdd8ff8\") " pod="openshift-console/console-6765c6547-hbtvp" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.417875 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88ed011c-aba3-49b4-9179-9bfc6fdd8ff8-trusted-ca-bundle\") pod \"console-6765c6547-hbtvp\" (UID: \"88ed011c-aba3-49b4-9179-9bfc6fdd8ff8\") " pod="openshift-console/console-6765c6547-hbtvp" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.417926 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/88ed011c-aba3-49b4-9179-9bfc6fdd8ff8-console-config\") pod \"console-6765c6547-hbtvp\" (UID: \"88ed011c-aba3-49b4-9179-9bfc6fdd8ff8\") " pod="openshift-console/console-6765c6547-hbtvp" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.417940 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/88ed011c-aba3-49b4-9179-9bfc6fdd8ff8-console-serving-cert\") pod \"console-6765c6547-hbtvp\" (UID: \"88ed011c-aba3-49b4-9179-9bfc6fdd8ff8\") " pod="openshift-console/console-6765c6547-hbtvp" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.417957 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/88ed011c-aba3-49b4-9179-9bfc6fdd8ff8-oauth-serving-cert\") pod \"console-6765c6547-hbtvp\" (UID: \"88ed011c-aba3-49b4-9179-9bfc6fdd8ff8\") " pod="openshift-console/console-6765c6547-hbtvp" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.418531 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/88ed011c-aba3-49b4-9179-9bfc6fdd8ff8-service-ca\") pod \"console-6765c6547-hbtvp\" (UID: \"88ed011c-aba3-49b4-9179-9bfc6fdd8ff8\") " pod="openshift-console/console-6765c6547-hbtvp" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.418857 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/88ed011c-aba3-49b4-9179-9bfc6fdd8ff8-oauth-serving-cert\") pod \"console-6765c6547-hbtvp\" (UID: \"88ed011c-aba3-49b4-9179-9bfc6fdd8ff8\") " pod="openshift-console/console-6765c6547-hbtvp" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.419280 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/88ed011c-aba3-49b4-9179-9bfc6fdd8ff8-console-config\") pod \"console-6765c6547-hbtvp\" (UID: \"88ed011c-aba3-49b4-9179-9bfc6fdd8ff8\") " pod="openshift-console/console-6765c6547-hbtvp" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.419564 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88ed011c-aba3-49b4-9179-9bfc6fdd8ff8-trusted-ca-bundle\") pod \"console-6765c6547-hbtvp\" (UID: \"88ed011c-aba3-49b4-9179-9bfc6fdd8ff8\") " pod="openshift-console/console-6765c6547-hbtvp" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.421048 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/88ed011c-aba3-49b4-9179-9bfc6fdd8ff8-console-serving-cert\") pod \"console-6765c6547-hbtvp\" (UID: \"88ed011c-aba3-49b4-9179-9bfc6fdd8ff8\") " pod="openshift-console/console-6765c6547-hbtvp" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.421492 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/88ed011c-aba3-49b4-9179-9bfc6fdd8ff8-console-oauth-config\") pod \"console-6765c6547-hbtvp\" (UID: \"88ed011c-aba3-49b4-9179-9bfc6fdd8ff8\") " pod="openshift-console/console-6765c6547-hbtvp" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.430876 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgndv\" (UniqueName: \"kubernetes.io/projected/88ed011c-aba3-49b4-9179-9bfc6fdd8ff8-kube-api-access-qgndv\") pod \"console-6765c6547-hbtvp\" (UID: \"88ed011c-aba3-49b4-9179-9bfc6fdd8ff8\") " pod="openshift-console/console-6765c6547-hbtvp" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.520610 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6765c6547-hbtvp" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.682991 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6765c6547-hbtvp"] Nov 23 06:54:18 crc kubenswrapper[4559]: W1123 06:54:18.690547 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88ed011c_aba3_49b4_9179_9bfc6fdd8ff8.slice/crio-d1c4b39eea43d58e3764d4914e3094a71f4c443892a8c825afc503ebc294eb44 WatchSource:0}: Error finding container d1c4b39eea43d58e3764d4914e3094a71f4c443892a8c825afc503ebc294eb44: Status 404 returned error can't find the container with id d1c4b39eea43d58e3764d4914e3094a71f4c443892a8c825afc503ebc294eb44 Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.823784 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b0662ab8-3035-4b7b-b615-15cc0b963af4-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-6qvx9\" (UID: \"b0662ab8-3035-4b7b-b615-15cc0b963af4\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6qvx9" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.828671 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b0662ab8-3035-4b7b-b615-15cc0b963af4-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-6qvx9\" (UID: \"b0662ab8-3035-4b7b-b615-15cc0b963af4\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6qvx9" Nov 23 06:54:18 crc kubenswrapper[4559]: I1123 06:54:18.931988 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6qvx9" Nov 23 06:54:19 crc kubenswrapper[4559]: E1123 06:54:19.012469 4559 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: failed to sync secret cache: timed out waiting for the condition Nov 23 06:54:19 crc kubenswrapper[4559]: E1123 06:54:19.012539 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6ff08b99-2dc5-4b9f-bef5-b0d34e76cb49-tls-key-pair podName:6ff08b99-2dc5-4b9f-bef5-b0d34e76cb49 nodeName:}" failed. No retries permitted until 2025-11-23 06:54:19.512523471 +0000 UTC m=+561.534509085 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/6ff08b99-2dc5-4b9f-bef5-b0d34e76cb49-tls-key-pair") pod "nmstate-webhook-6b89b748d8-hv2js" (UID: "6ff08b99-2dc5-4b9f-bef5-b0d34e76cb49") : failed to sync secret cache: timed out waiting for the condition Nov 23 06:54:19 crc kubenswrapper[4559]: I1123 06:54:19.052092 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6765c6547-hbtvp" event={"ID":"88ed011c-aba3-49b4-9179-9bfc6fdd8ff8","Type":"ContainerStarted","Data":"aa69d190a67e24f0b66c6b61e7040dd543c2db90f5da76728b5fc7265dbfeeb3"} Nov 23 06:54:19 crc kubenswrapper[4559]: I1123 06:54:19.052146 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6765c6547-hbtvp" event={"ID":"88ed011c-aba3-49b4-9179-9bfc6fdd8ff8","Type":"ContainerStarted","Data":"d1c4b39eea43d58e3764d4914e3094a71f4c443892a8c825afc503ebc294eb44"} Nov 23 06:54:19 crc kubenswrapper[4559]: I1123 06:54:19.053470 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-rpd7c" event={"ID":"3755e847-69fa-47e4-93f9-fe15df377011","Type":"ContainerStarted","Data":"8ecf552b7491d1fa3e0d6756b7fcca89c6b50b412f6bad392be072baacad8454"} Nov 23 06:54:19 crc kubenswrapper[4559]: I1123 06:54:19.054223 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-fglqj" event={"ID":"092f86a7-8e24-4a70-9af6-a4265c3688e5","Type":"ContainerStarted","Data":"289f0211bf56c1b16425b6a688c37dc85e8e1beba9c43778651a44f2266a0c05"} Nov 23 06:54:19 crc kubenswrapper[4559]: I1123 06:54:19.064665 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 23 06:54:19 crc kubenswrapper[4559]: I1123 06:54:19.068291 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6765c6547-hbtvp" podStartSLOduration=1.068277522 podStartE2EDuration="1.068277522s" podCreationTimestamp="2025-11-23 06:54:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:54:19.066881584 +0000 UTC m=+561.088867187" watchObservedRunningTime="2025-11-23 06:54:19.068277522 +0000 UTC m=+561.090263136" Nov 23 06:54:19 crc kubenswrapper[4559]: I1123 06:54:19.275842 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6qvx9"] Nov 23 06:54:19 crc kubenswrapper[4559]: W1123 06:54:19.283111 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb0662ab8_3035_4b7b_b615_15cc0b963af4.slice/crio-cf2e5568912dfe4b39664076f8e33baa27b71c6668341c3655dcd7d6778da45e WatchSource:0}: Error finding container cf2e5568912dfe4b39664076f8e33baa27b71c6668341c3655dcd7d6778da45e: Status 404 returned error can't find the container with id cf2e5568912dfe4b39664076f8e33baa27b71c6668341c3655dcd7d6778da45e Nov 23 06:54:19 crc kubenswrapper[4559]: I1123 06:54:19.531069 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/6ff08b99-2dc5-4b9f-bef5-b0d34e76cb49-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-hv2js\" (UID: \"6ff08b99-2dc5-4b9f-bef5-b0d34e76cb49\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-hv2js" Nov 23 06:54:19 crc kubenswrapper[4559]: I1123 06:54:19.536426 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/6ff08b99-2dc5-4b9f-bef5-b0d34e76cb49-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-hv2js\" (UID: \"6ff08b99-2dc5-4b9f-bef5-b0d34e76cb49\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-hv2js" Nov 23 06:54:19 crc kubenswrapper[4559]: I1123 06:54:19.653875 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-hv2js" Nov 23 06:54:19 crc kubenswrapper[4559]: I1123 06:54:19.817346 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-hv2js"] Nov 23 06:54:19 crc kubenswrapper[4559]: W1123 06:54:19.820436 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ff08b99_2dc5_4b9f_bef5_b0d34e76cb49.slice/crio-ca4b24fe5aa712863571c194afa4d137ecb198c809086af650c6d51aba78ac3c WatchSource:0}: Error finding container ca4b24fe5aa712863571c194afa4d137ecb198c809086af650c6d51aba78ac3c: Status 404 returned error can't find the container with id ca4b24fe5aa712863571c194afa4d137ecb198c809086af650c6d51aba78ac3c Nov 23 06:54:20 crc kubenswrapper[4559]: I1123 06:54:20.060079 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-hv2js" event={"ID":"6ff08b99-2dc5-4b9f-bef5-b0d34e76cb49","Type":"ContainerStarted","Data":"ca4b24fe5aa712863571c194afa4d137ecb198c809086af650c6d51aba78ac3c"} Nov 23 06:54:20 crc kubenswrapper[4559]: I1123 06:54:20.061758 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6qvx9" event={"ID":"b0662ab8-3035-4b7b-b615-15cc0b963af4","Type":"ContainerStarted","Data":"cf2e5568912dfe4b39664076f8e33baa27b71c6668341c3655dcd7d6778da45e"} Nov 23 06:54:21 crc kubenswrapper[4559]: I1123 06:54:21.067486 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-hv2js" event={"ID":"6ff08b99-2dc5-4b9f-bef5-b0d34e76cb49","Type":"ContainerStarted","Data":"cf5cf72c6ceaf7257726e225d0c142321e4bd8d78ef091b725372e2ac430954e"} Nov 23 06:54:21 crc kubenswrapper[4559]: I1123 06:54:21.067940 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-hv2js" Nov 23 06:54:21 crc kubenswrapper[4559]: I1123 06:54:21.069121 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-fglqj" event={"ID":"092f86a7-8e24-4a70-9af6-a4265c3688e5","Type":"ContainerStarted","Data":"cad9a967273752099e4e02d893c98faf1152c6d0a6ab6cbaccb011fdd04d88fb"} Nov 23 06:54:21 crc kubenswrapper[4559]: I1123 06:54:21.070072 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-fglqj" Nov 23 06:54:21 crc kubenswrapper[4559]: I1123 06:54:21.071719 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-rpd7c" event={"ID":"3755e847-69fa-47e4-93f9-fe15df377011","Type":"ContainerStarted","Data":"29f615b13bd31a4c7d4e8f3106c3d0e48b70fa62a25d23eae7086ea8f495e62e"} Nov 23 06:54:21 crc kubenswrapper[4559]: I1123 06:54:21.080398 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-hv2js" podStartSLOduration=3.322222794 podStartE2EDuration="4.080379699s" podCreationTimestamp="2025-11-23 06:54:17 +0000 UTC" firstStartedPulling="2025-11-23 06:54:19.822274662 +0000 UTC m=+561.844260276" lastFinishedPulling="2025-11-23 06:54:20.580431566 +0000 UTC m=+562.602417181" observedRunningTime="2025-11-23 06:54:21.078564622 +0000 UTC m=+563.100550247" watchObservedRunningTime="2025-11-23 06:54:21.080379699 +0000 UTC m=+563.102365304" Nov 23 06:54:21 crc kubenswrapper[4559]: I1123 06:54:21.093456 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-fglqj" podStartSLOduration=1.8381791509999998 podStartE2EDuration="4.093430588s" podCreationTimestamp="2025-11-23 06:54:17 +0000 UTC" firstStartedPulling="2025-11-23 06:54:18.212002805 +0000 UTC m=+560.233988419" lastFinishedPulling="2025-11-23 06:54:20.467254241 +0000 UTC m=+562.489239856" observedRunningTime="2025-11-23 06:54:21.09204571 +0000 UTC m=+563.114031325" watchObservedRunningTime="2025-11-23 06:54:21.093430588 +0000 UTC m=+563.115416202" Nov 23 06:54:22 crc kubenswrapper[4559]: I1123 06:54:22.082770 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6qvx9" event={"ID":"b0662ab8-3035-4b7b-b615-15cc0b963af4","Type":"ContainerStarted","Data":"e76c60f317f40cb1af8f1b77711f833cde7d891cf394c914087a143e223dab66"} Nov 23 06:54:22 crc kubenswrapper[4559]: I1123 06:54:22.100533 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6qvx9" podStartSLOduration=2.093877598 podStartE2EDuration="4.100515985s" podCreationTimestamp="2025-11-23 06:54:18 +0000 UTC" firstStartedPulling="2025-11-23 06:54:19.286093441 +0000 UTC m=+561.308079056" lastFinishedPulling="2025-11-23 06:54:21.292731829 +0000 UTC m=+563.314717443" observedRunningTime="2025-11-23 06:54:22.095470213 +0000 UTC m=+564.117455827" watchObservedRunningTime="2025-11-23 06:54:22.100515985 +0000 UTC m=+564.122501599" Nov 23 06:54:23 crc kubenswrapper[4559]: I1123 06:54:23.090570 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-rpd7c" event={"ID":"3755e847-69fa-47e4-93f9-fe15df377011","Type":"ContainerStarted","Data":"6cbc23b075c1782528a24addee4758ff9aa3b4d55aa235b2358ba7ecaf1f7f43"} Nov 23 06:54:26 crc kubenswrapper[4559]: I1123 06:54:26.167498 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 06:54:26 crc kubenswrapper[4559]: I1123 06:54:26.167923 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 06:54:26 crc kubenswrapper[4559]: I1123 06:54:26.167968 4559 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" Nov 23 06:54:26 crc kubenswrapper[4559]: I1123 06:54:26.168497 4559 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"caa8c42bf6f598aae815f53f4281cd35582267034e69906ef9d6b656540ea6ab"} pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 06:54:26 crc kubenswrapper[4559]: I1123 06:54:26.168554 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" containerID="cri-o://caa8c42bf6f598aae815f53f4281cd35582267034e69906ef9d6b656540ea6ab" gracePeriod=600 Nov 23 06:54:27 crc kubenswrapper[4559]: I1123 06:54:27.113397 4559 generic.go:334] "Generic (PLEG): container finished" podID="4731beee-0cac-4189-8a70-743b0b709095" containerID="caa8c42bf6f598aae815f53f4281cd35582267034e69906ef9d6b656540ea6ab" exitCode=0 Nov 23 06:54:27 crc kubenswrapper[4559]: I1123 06:54:27.113477 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerDied","Data":"caa8c42bf6f598aae815f53f4281cd35582267034e69906ef9d6b656540ea6ab"} Nov 23 06:54:27 crc kubenswrapper[4559]: I1123 06:54:27.113938 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerStarted","Data":"c1a47810490d41ffdaf229d9abf61dc047a9840ab17302f7f8ed9bcafa8fe6db"} Nov 23 06:54:27 crc kubenswrapper[4559]: I1123 06:54:27.113973 4559 scope.go:117] "RemoveContainer" containerID="232a78e72da23760d36189f8c69ea7f78803b94ef0a735cc1b9c323800c5683a" Nov 23 06:54:27 crc kubenswrapper[4559]: I1123 06:54:27.129634 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-rpd7c" podStartSLOduration=5.874793057 podStartE2EDuration="10.12959903s" podCreationTimestamp="2025-11-23 06:54:17 +0000 UTC" firstStartedPulling="2025-11-23 06:54:18.39415024 +0000 UTC m=+560.416135854" lastFinishedPulling="2025-11-23 06:54:22.648956213 +0000 UTC m=+564.670941827" observedRunningTime="2025-11-23 06:54:23.119703804 +0000 UTC m=+565.141689429" watchObservedRunningTime="2025-11-23 06:54:27.12959903 +0000 UTC m=+569.151584644" Nov 23 06:54:28 crc kubenswrapper[4559]: I1123 06:54:28.194715 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-fglqj" Nov 23 06:54:28 crc kubenswrapper[4559]: I1123 06:54:28.521495 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6765c6547-hbtvp" Nov 23 06:54:28 crc kubenswrapper[4559]: I1123 06:54:28.521561 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6765c6547-hbtvp" Nov 23 06:54:28 crc kubenswrapper[4559]: I1123 06:54:28.527410 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6765c6547-hbtvp" Nov 23 06:54:29 crc kubenswrapper[4559]: I1123 06:54:29.130080 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6765c6547-hbtvp" Nov 23 06:54:29 crc kubenswrapper[4559]: I1123 06:54:29.177263 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-xvd4d"] Nov 23 06:54:39 crc kubenswrapper[4559]: I1123 06:54:39.658985 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-hv2js" Nov 23 06:54:49 crc kubenswrapper[4559]: I1123 06:54:49.137400 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458"] Nov 23 06:54:49 crc kubenswrapper[4559]: I1123 06:54:49.138807 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458" Nov 23 06:54:49 crc kubenswrapper[4559]: I1123 06:54:49.141077 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 23 06:54:49 crc kubenswrapper[4559]: I1123 06:54:49.143998 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458"] Nov 23 06:54:49 crc kubenswrapper[4559]: I1123 06:54:49.184479 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/23f858ff-40d3-4f2a-ac00-f7e3b8c544f1-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458\" (UID: \"23f858ff-40d3-4f2a-ac00-f7e3b8c544f1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458" Nov 23 06:54:49 crc kubenswrapper[4559]: I1123 06:54:49.184523 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsrdx\" (UniqueName: \"kubernetes.io/projected/23f858ff-40d3-4f2a-ac00-f7e3b8c544f1-kube-api-access-dsrdx\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458\" (UID: \"23f858ff-40d3-4f2a-ac00-f7e3b8c544f1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458" Nov 23 06:54:49 crc kubenswrapper[4559]: I1123 06:54:49.184561 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/23f858ff-40d3-4f2a-ac00-f7e3b8c544f1-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458\" (UID: \"23f858ff-40d3-4f2a-ac00-f7e3b8c544f1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458" Nov 23 06:54:49 crc kubenswrapper[4559]: I1123 06:54:49.285758 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsrdx\" (UniqueName: \"kubernetes.io/projected/23f858ff-40d3-4f2a-ac00-f7e3b8c544f1-kube-api-access-dsrdx\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458\" (UID: \"23f858ff-40d3-4f2a-ac00-f7e3b8c544f1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458" Nov 23 06:54:49 crc kubenswrapper[4559]: I1123 06:54:49.285844 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/23f858ff-40d3-4f2a-ac00-f7e3b8c544f1-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458\" (UID: \"23f858ff-40d3-4f2a-ac00-f7e3b8c544f1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458" Nov 23 06:54:49 crc kubenswrapper[4559]: I1123 06:54:49.285944 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/23f858ff-40d3-4f2a-ac00-f7e3b8c544f1-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458\" (UID: \"23f858ff-40d3-4f2a-ac00-f7e3b8c544f1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458" Nov 23 06:54:49 crc kubenswrapper[4559]: I1123 06:54:49.286302 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/23f858ff-40d3-4f2a-ac00-f7e3b8c544f1-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458\" (UID: \"23f858ff-40d3-4f2a-ac00-f7e3b8c544f1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458" Nov 23 06:54:49 crc kubenswrapper[4559]: I1123 06:54:49.286339 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/23f858ff-40d3-4f2a-ac00-f7e3b8c544f1-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458\" (UID: \"23f858ff-40d3-4f2a-ac00-f7e3b8c544f1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458" Nov 23 06:54:49 crc kubenswrapper[4559]: I1123 06:54:49.304327 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsrdx\" (UniqueName: \"kubernetes.io/projected/23f858ff-40d3-4f2a-ac00-f7e3b8c544f1-kube-api-access-dsrdx\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458\" (UID: \"23f858ff-40d3-4f2a-ac00-f7e3b8c544f1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458" Nov 23 06:54:49 crc kubenswrapper[4559]: I1123 06:54:49.454408 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458" Nov 23 06:54:49 crc kubenswrapper[4559]: I1123 06:54:49.815130 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458"] Nov 23 06:54:50 crc kubenswrapper[4559]: I1123 06:54:50.245889 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458" event={"ID":"23f858ff-40d3-4f2a-ac00-f7e3b8c544f1","Type":"ContainerStarted","Data":"d7096192d23ef3cd5bacb6e814a3ab725d600b75714da03798c066f66fc8c50d"} Nov 23 06:54:50 crc kubenswrapper[4559]: I1123 06:54:50.246126 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458" event={"ID":"23f858ff-40d3-4f2a-ac00-f7e3b8c544f1","Type":"ContainerStarted","Data":"8f7ad0a5f4771ad9cd016fbac55977bdef1a4069fd190110c80c168890e96fad"} Nov 23 06:54:51 crc kubenswrapper[4559]: I1123 06:54:51.252429 4559 generic.go:334] "Generic (PLEG): container finished" podID="23f858ff-40d3-4f2a-ac00-f7e3b8c544f1" containerID="d7096192d23ef3cd5bacb6e814a3ab725d600b75714da03798c066f66fc8c50d" exitCode=0 Nov 23 06:54:51 crc kubenswrapper[4559]: I1123 06:54:51.252464 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458" event={"ID":"23f858ff-40d3-4f2a-ac00-f7e3b8c544f1","Type":"ContainerDied","Data":"d7096192d23ef3cd5bacb6e814a3ab725d600b75714da03798c066f66fc8c50d"} Nov 23 06:54:53 crc kubenswrapper[4559]: I1123 06:54:53.265059 4559 generic.go:334] "Generic (PLEG): container finished" podID="23f858ff-40d3-4f2a-ac00-f7e3b8c544f1" containerID="9647e82f1a956495b949c827be2338260750e2c38c5da82aade124868c466bfb" exitCode=0 Nov 23 06:54:53 crc kubenswrapper[4559]: I1123 06:54:53.265121 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458" event={"ID":"23f858ff-40d3-4f2a-ac00-f7e3b8c544f1","Type":"ContainerDied","Data":"9647e82f1a956495b949c827be2338260750e2c38c5da82aade124868c466bfb"} Nov 23 06:54:54 crc kubenswrapper[4559]: I1123 06:54:54.205266 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-xvd4d" podUID="6796a9bd-a428-4b0a-8723-d8a6a252de84" containerName="console" containerID="cri-o://b62a236308102438a80cc28535cfdc1e01fd091cc74122207037decdbb62735c" gracePeriod=15 Nov 23 06:54:54 crc kubenswrapper[4559]: I1123 06:54:54.274130 4559 generic.go:334] "Generic (PLEG): container finished" podID="23f858ff-40d3-4f2a-ac00-f7e3b8c544f1" containerID="8b1279f5aa090b4beee9fdb9d3d3c5fdd6e3c36b78b16022a3afc7d2cf4c785d" exitCode=0 Nov 23 06:54:54 crc kubenswrapper[4559]: I1123 06:54:54.278318 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458" event={"ID":"23f858ff-40d3-4f2a-ac00-f7e3b8c544f1","Type":"ContainerDied","Data":"8b1279f5aa090b4beee9fdb9d3d3c5fdd6e3c36b78b16022a3afc7d2cf4c785d"} Nov 23 06:54:54 crc kubenswrapper[4559]: I1123 06:54:54.507495 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-xvd4d_6796a9bd-a428-4b0a-8723-d8a6a252de84/console/0.log" Nov 23 06:54:54 crc kubenswrapper[4559]: I1123 06:54:54.507569 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:54:54 crc kubenswrapper[4559]: I1123 06:54:54.548789 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6796a9bd-a428-4b0a-8723-d8a6a252de84-console-oauth-config\") pod \"6796a9bd-a428-4b0a-8723-d8a6a252de84\" (UID: \"6796a9bd-a428-4b0a-8723-d8a6a252de84\") " Nov 23 06:54:54 crc kubenswrapper[4559]: I1123 06:54:54.548839 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6796a9bd-a428-4b0a-8723-d8a6a252de84-oauth-serving-cert\") pod \"6796a9bd-a428-4b0a-8723-d8a6a252de84\" (UID: \"6796a9bd-a428-4b0a-8723-d8a6a252de84\") " Nov 23 06:54:54 crc kubenswrapper[4559]: I1123 06:54:54.548917 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fjff\" (UniqueName: \"kubernetes.io/projected/6796a9bd-a428-4b0a-8723-d8a6a252de84-kube-api-access-8fjff\") pod \"6796a9bd-a428-4b0a-8723-d8a6a252de84\" (UID: \"6796a9bd-a428-4b0a-8723-d8a6a252de84\") " Nov 23 06:54:54 crc kubenswrapper[4559]: I1123 06:54:54.548961 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6796a9bd-a428-4b0a-8723-d8a6a252de84-service-ca\") pod \"6796a9bd-a428-4b0a-8723-d8a6a252de84\" (UID: \"6796a9bd-a428-4b0a-8723-d8a6a252de84\") " Nov 23 06:54:54 crc kubenswrapper[4559]: I1123 06:54:54.549047 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6796a9bd-a428-4b0a-8723-d8a6a252de84-console-serving-cert\") pod \"6796a9bd-a428-4b0a-8723-d8a6a252de84\" (UID: \"6796a9bd-a428-4b0a-8723-d8a6a252de84\") " Nov 23 06:54:54 crc kubenswrapper[4559]: I1123 06:54:54.549065 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6796a9bd-a428-4b0a-8723-d8a6a252de84-console-config\") pod \"6796a9bd-a428-4b0a-8723-d8a6a252de84\" (UID: \"6796a9bd-a428-4b0a-8723-d8a6a252de84\") " Nov 23 06:54:54 crc kubenswrapper[4559]: I1123 06:54:54.549095 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6796a9bd-a428-4b0a-8723-d8a6a252de84-trusted-ca-bundle\") pod \"6796a9bd-a428-4b0a-8723-d8a6a252de84\" (UID: \"6796a9bd-a428-4b0a-8723-d8a6a252de84\") " Nov 23 06:54:54 crc kubenswrapper[4559]: I1123 06:54:54.549790 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6796a9bd-a428-4b0a-8723-d8a6a252de84-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6796a9bd-a428-4b0a-8723-d8a6a252de84" (UID: "6796a9bd-a428-4b0a-8723-d8a6a252de84"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:54:54 crc kubenswrapper[4559]: I1123 06:54:54.549805 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6796a9bd-a428-4b0a-8723-d8a6a252de84-service-ca" (OuterVolumeSpecName: "service-ca") pod "6796a9bd-a428-4b0a-8723-d8a6a252de84" (UID: "6796a9bd-a428-4b0a-8723-d8a6a252de84"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:54:54 crc kubenswrapper[4559]: I1123 06:54:54.549812 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6796a9bd-a428-4b0a-8723-d8a6a252de84-console-config" (OuterVolumeSpecName: "console-config") pod "6796a9bd-a428-4b0a-8723-d8a6a252de84" (UID: "6796a9bd-a428-4b0a-8723-d8a6a252de84"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:54:54 crc kubenswrapper[4559]: I1123 06:54:54.549966 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6796a9bd-a428-4b0a-8723-d8a6a252de84-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "6796a9bd-a428-4b0a-8723-d8a6a252de84" (UID: "6796a9bd-a428-4b0a-8723-d8a6a252de84"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:54:54 crc kubenswrapper[4559]: I1123 06:54:54.554926 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6796a9bd-a428-4b0a-8723-d8a6a252de84-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "6796a9bd-a428-4b0a-8723-d8a6a252de84" (UID: "6796a9bd-a428-4b0a-8723-d8a6a252de84"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:54:54 crc kubenswrapper[4559]: I1123 06:54:54.555627 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6796a9bd-a428-4b0a-8723-d8a6a252de84-kube-api-access-8fjff" (OuterVolumeSpecName: "kube-api-access-8fjff") pod "6796a9bd-a428-4b0a-8723-d8a6a252de84" (UID: "6796a9bd-a428-4b0a-8723-d8a6a252de84"). InnerVolumeSpecName "kube-api-access-8fjff". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:54:54 crc kubenswrapper[4559]: I1123 06:54:54.555680 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6796a9bd-a428-4b0a-8723-d8a6a252de84-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "6796a9bd-a428-4b0a-8723-d8a6a252de84" (UID: "6796a9bd-a428-4b0a-8723-d8a6a252de84"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:54:54 crc kubenswrapper[4559]: I1123 06:54:54.650018 4559 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6796a9bd-a428-4b0a-8723-d8a6a252de84-service-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:54:54 crc kubenswrapper[4559]: I1123 06:54:54.650061 4559 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6796a9bd-a428-4b0a-8723-d8a6a252de84-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:54:54 crc kubenswrapper[4559]: I1123 06:54:54.650074 4559 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6796a9bd-a428-4b0a-8723-d8a6a252de84-console-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:54:54 crc kubenswrapper[4559]: I1123 06:54:54.650087 4559 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6796a9bd-a428-4b0a-8723-d8a6a252de84-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:54:54 crc kubenswrapper[4559]: I1123 06:54:54.650100 4559 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6796a9bd-a428-4b0a-8723-d8a6a252de84-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:54:54 crc kubenswrapper[4559]: I1123 06:54:54.650112 4559 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6796a9bd-a428-4b0a-8723-d8a6a252de84-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:54:54 crc kubenswrapper[4559]: I1123 06:54:54.650123 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fjff\" (UniqueName: \"kubernetes.io/projected/6796a9bd-a428-4b0a-8723-d8a6a252de84-kube-api-access-8fjff\") on node \"crc\" DevicePath \"\"" Nov 23 06:54:55 crc kubenswrapper[4559]: I1123 06:54:55.280775 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-xvd4d_6796a9bd-a428-4b0a-8723-d8a6a252de84/console/0.log" Nov 23 06:54:55 crc kubenswrapper[4559]: I1123 06:54:55.280826 4559 generic.go:334] "Generic (PLEG): container finished" podID="6796a9bd-a428-4b0a-8723-d8a6a252de84" containerID="b62a236308102438a80cc28535cfdc1e01fd091cc74122207037decdbb62735c" exitCode=2 Nov 23 06:54:55 crc kubenswrapper[4559]: I1123 06:54:55.280895 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-xvd4d" event={"ID":"6796a9bd-a428-4b0a-8723-d8a6a252de84","Type":"ContainerDied","Data":"b62a236308102438a80cc28535cfdc1e01fd091cc74122207037decdbb62735c"} Nov 23 06:54:55 crc kubenswrapper[4559]: I1123 06:54:55.280955 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-xvd4d" event={"ID":"6796a9bd-a428-4b0a-8723-d8a6a252de84","Type":"ContainerDied","Data":"e813441ff551059a86acca42beb61e7154ff1ea903a964f49529e56b8afc01fc"} Nov 23 06:54:55 crc kubenswrapper[4559]: I1123 06:54:55.280978 4559 scope.go:117] "RemoveContainer" containerID="b62a236308102438a80cc28535cfdc1e01fd091cc74122207037decdbb62735c" Nov 23 06:54:55 crc kubenswrapper[4559]: I1123 06:54:55.280915 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-xvd4d" Nov 23 06:54:55 crc kubenswrapper[4559]: I1123 06:54:55.300786 4559 scope.go:117] "RemoveContainer" containerID="b62a236308102438a80cc28535cfdc1e01fd091cc74122207037decdbb62735c" Nov 23 06:54:55 crc kubenswrapper[4559]: E1123 06:54:55.302580 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b62a236308102438a80cc28535cfdc1e01fd091cc74122207037decdbb62735c\": container with ID starting with b62a236308102438a80cc28535cfdc1e01fd091cc74122207037decdbb62735c not found: ID does not exist" containerID="b62a236308102438a80cc28535cfdc1e01fd091cc74122207037decdbb62735c" Nov 23 06:54:55 crc kubenswrapper[4559]: I1123 06:54:55.302628 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b62a236308102438a80cc28535cfdc1e01fd091cc74122207037decdbb62735c"} err="failed to get container status \"b62a236308102438a80cc28535cfdc1e01fd091cc74122207037decdbb62735c\": rpc error: code = NotFound desc = could not find container \"b62a236308102438a80cc28535cfdc1e01fd091cc74122207037decdbb62735c\": container with ID starting with b62a236308102438a80cc28535cfdc1e01fd091cc74122207037decdbb62735c not found: ID does not exist" Nov 23 06:54:55 crc kubenswrapper[4559]: I1123 06:54:55.308223 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-xvd4d"] Nov 23 06:54:55 crc kubenswrapper[4559]: I1123 06:54:55.312467 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-xvd4d"] Nov 23 06:54:55 crc kubenswrapper[4559]: I1123 06:54:55.466189 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458" Nov 23 06:54:55 crc kubenswrapper[4559]: I1123 06:54:55.661804 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/23f858ff-40d3-4f2a-ac00-f7e3b8c544f1-util\") pod \"23f858ff-40d3-4f2a-ac00-f7e3b8c544f1\" (UID: \"23f858ff-40d3-4f2a-ac00-f7e3b8c544f1\") " Nov 23 06:54:55 crc kubenswrapper[4559]: I1123 06:54:55.661914 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/23f858ff-40d3-4f2a-ac00-f7e3b8c544f1-bundle\") pod \"23f858ff-40d3-4f2a-ac00-f7e3b8c544f1\" (UID: \"23f858ff-40d3-4f2a-ac00-f7e3b8c544f1\") " Nov 23 06:54:55 crc kubenswrapper[4559]: I1123 06:54:55.661982 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsrdx\" (UniqueName: \"kubernetes.io/projected/23f858ff-40d3-4f2a-ac00-f7e3b8c544f1-kube-api-access-dsrdx\") pod \"23f858ff-40d3-4f2a-ac00-f7e3b8c544f1\" (UID: \"23f858ff-40d3-4f2a-ac00-f7e3b8c544f1\") " Nov 23 06:54:55 crc kubenswrapper[4559]: I1123 06:54:55.662935 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23f858ff-40d3-4f2a-ac00-f7e3b8c544f1-bundle" (OuterVolumeSpecName: "bundle") pod "23f858ff-40d3-4f2a-ac00-f7e3b8c544f1" (UID: "23f858ff-40d3-4f2a-ac00-f7e3b8c544f1"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:54:55 crc kubenswrapper[4559]: I1123 06:54:55.667768 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23f858ff-40d3-4f2a-ac00-f7e3b8c544f1-kube-api-access-dsrdx" (OuterVolumeSpecName: "kube-api-access-dsrdx") pod "23f858ff-40d3-4f2a-ac00-f7e3b8c544f1" (UID: "23f858ff-40d3-4f2a-ac00-f7e3b8c544f1"). InnerVolumeSpecName "kube-api-access-dsrdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:54:55 crc kubenswrapper[4559]: I1123 06:54:55.703533 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23f858ff-40d3-4f2a-ac00-f7e3b8c544f1-util" (OuterVolumeSpecName: "util") pod "23f858ff-40d3-4f2a-ac00-f7e3b8c544f1" (UID: "23f858ff-40d3-4f2a-ac00-f7e3b8c544f1"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:54:55 crc kubenswrapper[4559]: I1123 06:54:55.762889 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsrdx\" (UniqueName: \"kubernetes.io/projected/23f858ff-40d3-4f2a-ac00-f7e3b8c544f1-kube-api-access-dsrdx\") on node \"crc\" DevicePath \"\"" Nov 23 06:54:55 crc kubenswrapper[4559]: I1123 06:54:55.762911 4559 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/23f858ff-40d3-4f2a-ac00-f7e3b8c544f1-util\") on node \"crc\" DevicePath \"\"" Nov 23 06:54:55 crc kubenswrapper[4559]: I1123 06:54:55.762919 4559 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/23f858ff-40d3-4f2a-ac00-f7e3b8c544f1-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:54:56 crc kubenswrapper[4559]: I1123 06:54:56.279684 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6796a9bd-a428-4b0a-8723-d8a6a252de84" path="/var/lib/kubelet/pods/6796a9bd-a428-4b0a-8723-d8a6a252de84/volumes" Nov 23 06:54:56 crc kubenswrapper[4559]: I1123 06:54:56.286891 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458" event={"ID":"23f858ff-40d3-4f2a-ac00-f7e3b8c544f1","Type":"ContainerDied","Data":"8f7ad0a5f4771ad9cd016fbac55977bdef1a4069fd190110c80c168890e96fad"} Nov 23 06:54:56 crc kubenswrapper[4559]: I1123 06:54:56.286941 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f7ad0a5f4771ad9cd016fbac55977bdef1a4069fd190110c80c168890e96fad" Nov 23 06:54:56 crc kubenswrapper[4559]: I1123 06:54:56.286913 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.337020 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5c8ccb79b7-ngqhc"] Nov 23 06:55:05 crc kubenswrapper[4559]: E1123 06:55:05.337834 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23f858ff-40d3-4f2a-ac00-f7e3b8c544f1" containerName="pull" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.337863 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="23f858ff-40d3-4f2a-ac00-f7e3b8c544f1" containerName="pull" Nov 23 06:55:05 crc kubenswrapper[4559]: E1123 06:55:05.337878 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23f858ff-40d3-4f2a-ac00-f7e3b8c544f1" containerName="extract" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.337884 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="23f858ff-40d3-4f2a-ac00-f7e3b8c544f1" containerName="extract" Nov 23 06:55:05 crc kubenswrapper[4559]: E1123 06:55:05.337891 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23f858ff-40d3-4f2a-ac00-f7e3b8c544f1" containerName="util" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.337897 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="23f858ff-40d3-4f2a-ac00-f7e3b8c544f1" containerName="util" Nov 23 06:55:05 crc kubenswrapper[4559]: E1123 06:55:05.337906 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6796a9bd-a428-4b0a-8723-d8a6a252de84" containerName="console" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.337911 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="6796a9bd-a428-4b0a-8723-d8a6a252de84" containerName="console" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.338021 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="23f858ff-40d3-4f2a-ac00-f7e3b8c544f1" containerName="extract" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.338033 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="6796a9bd-a428-4b0a-8723-d8a6a252de84" containerName="console" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.338467 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5c8ccb79b7-ngqhc" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.339952 4559 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.340578 4559 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-xvmtc" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.340908 4559 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.341021 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.341016 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.358850 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5c8ccb79b7-ngqhc"] Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.374949 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e57be21d-6409-4c23-b745-ec26bdf7e98d-webhook-cert\") pod \"metallb-operator-controller-manager-5c8ccb79b7-ngqhc\" (UID: \"e57be21d-6409-4c23-b745-ec26bdf7e98d\") " pod="metallb-system/metallb-operator-controller-manager-5c8ccb79b7-ngqhc" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.375063 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rprch\" (UniqueName: \"kubernetes.io/projected/e57be21d-6409-4c23-b745-ec26bdf7e98d-kube-api-access-rprch\") pod \"metallb-operator-controller-manager-5c8ccb79b7-ngqhc\" (UID: \"e57be21d-6409-4c23-b745-ec26bdf7e98d\") " pod="metallb-system/metallb-operator-controller-manager-5c8ccb79b7-ngqhc" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.375158 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e57be21d-6409-4c23-b745-ec26bdf7e98d-apiservice-cert\") pod \"metallb-operator-controller-manager-5c8ccb79b7-ngqhc\" (UID: \"e57be21d-6409-4c23-b745-ec26bdf7e98d\") " pod="metallb-system/metallb-operator-controller-manager-5c8ccb79b7-ngqhc" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.476290 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e57be21d-6409-4c23-b745-ec26bdf7e98d-webhook-cert\") pod \"metallb-operator-controller-manager-5c8ccb79b7-ngqhc\" (UID: \"e57be21d-6409-4c23-b745-ec26bdf7e98d\") " pod="metallb-system/metallb-operator-controller-manager-5c8ccb79b7-ngqhc" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.476347 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rprch\" (UniqueName: \"kubernetes.io/projected/e57be21d-6409-4c23-b745-ec26bdf7e98d-kube-api-access-rprch\") pod \"metallb-operator-controller-manager-5c8ccb79b7-ngqhc\" (UID: \"e57be21d-6409-4c23-b745-ec26bdf7e98d\") " pod="metallb-system/metallb-operator-controller-manager-5c8ccb79b7-ngqhc" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.476380 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e57be21d-6409-4c23-b745-ec26bdf7e98d-apiservice-cert\") pod \"metallb-operator-controller-manager-5c8ccb79b7-ngqhc\" (UID: \"e57be21d-6409-4c23-b745-ec26bdf7e98d\") " pod="metallb-system/metallb-operator-controller-manager-5c8ccb79b7-ngqhc" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.482460 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e57be21d-6409-4c23-b745-ec26bdf7e98d-webhook-cert\") pod \"metallb-operator-controller-manager-5c8ccb79b7-ngqhc\" (UID: \"e57be21d-6409-4c23-b745-ec26bdf7e98d\") " pod="metallb-system/metallb-operator-controller-manager-5c8ccb79b7-ngqhc" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.483043 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e57be21d-6409-4c23-b745-ec26bdf7e98d-apiservice-cert\") pod \"metallb-operator-controller-manager-5c8ccb79b7-ngqhc\" (UID: \"e57be21d-6409-4c23-b745-ec26bdf7e98d\") " pod="metallb-system/metallb-operator-controller-manager-5c8ccb79b7-ngqhc" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.495971 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rprch\" (UniqueName: \"kubernetes.io/projected/e57be21d-6409-4c23-b745-ec26bdf7e98d-kube-api-access-rprch\") pod \"metallb-operator-controller-manager-5c8ccb79b7-ngqhc\" (UID: \"e57be21d-6409-4c23-b745-ec26bdf7e98d\") " pod="metallb-system/metallb-operator-controller-manager-5c8ccb79b7-ngqhc" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.654544 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5c8ccb79b7-ngqhc" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.682125 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-6bc96cc64c-zfqmv"] Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.682995 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6bc96cc64c-zfqmv" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.686413 4559 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.686593 4559 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-l22wj" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.686665 4559 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.707463 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6bc96cc64c-zfqmv"] Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.781446 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/85db5798-c470-4ec9-91c3-c9bdd9db327d-webhook-cert\") pod \"metallb-operator-webhook-server-6bc96cc64c-zfqmv\" (UID: \"85db5798-c470-4ec9-91c3-c9bdd9db327d\") " pod="metallb-system/metallb-operator-webhook-server-6bc96cc64c-zfqmv" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.781783 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/85db5798-c470-4ec9-91c3-c9bdd9db327d-apiservice-cert\") pod \"metallb-operator-webhook-server-6bc96cc64c-zfqmv\" (UID: \"85db5798-c470-4ec9-91c3-c9bdd9db327d\") " pod="metallb-system/metallb-operator-webhook-server-6bc96cc64c-zfqmv" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.781845 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcr68\" (UniqueName: \"kubernetes.io/projected/85db5798-c470-4ec9-91c3-c9bdd9db327d-kube-api-access-tcr68\") pod \"metallb-operator-webhook-server-6bc96cc64c-zfqmv\" (UID: \"85db5798-c470-4ec9-91c3-c9bdd9db327d\") " pod="metallb-system/metallb-operator-webhook-server-6bc96cc64c-zfqmv" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.883143 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/85db5798-c470-4ec9-91c3-c9bdd9db327d-apiservice-cert\") pod \"metallb-operator-webhook-server-6bc96cc64c-zfqmv\" (UID: \"85db5798-c470-4ec9-91c3-c9bdd9db327d\") " pod="metallb-system/metallb-operator-webhook-server-6bc96cc64c-zfqmv" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.883220 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcr68\" (UniqueName: \"kubernetes.io/projected/85db5798-c470-4ec9-91c3-c9bdd9db327d-kube-api-access-tcr68\") pod \"metallb-operator-webhook-server-6bc96cc64c-zfqmv\" (UID: \"85db5798-c470-4ec9-91c3-c9bdd9db327d\") " pod="metallb-system/metallb-operator-webhook-server-6bc96cc64c-zfqmv" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.883264 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/85db5798-c470-4ec9-91c3-c9bdd9db327d-webhook-cert\") pod \"metallb-operator-webhook-server-6bc96cc64c-zfqmv\" (UID: \"85db5798-c470-4ec9-91c3-c9bdd9db327d\") " pod="metallb-system/metallb-operator-webhook-server-6bc96cc64c-zfqmv" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.887245 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/85db5798-c470-4ec9-91c3-c9bdd9db327d-apiservice-cert\") pod \"metallb-operator-webhook-server-6bc96cc64c-zfqmv\" (UID: \"85db5798-c470-4ec9-91c3-c9bdd9db327d\") " pod="metallb-system/metallb-operator-webhook-server-6bc96cc64c-zfqmv" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.888591 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/85db5798-c470-4ec9-91c3-c9bdd9db327d-webhook-cert\") pod \"metallb-operator-webhook-server-6bc96cc64c-zfqmv\" (UID: \"85db5798-c470-4ec9-91c3-c9bdd9db327d\") " pod="metallb-system/metallb-operator-webhook-server-6bc96cc64c-zfqmv" Nov 23 06:55:05 crc kubenswrapper[4559]: I1123 06:55:05.898757 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcr68\" (UniqueName: \"kubernetes.io/projected/85db5798-c470-4ec9-91c3-c9bdd9db327d-kube-api-access-tcr68\") pod \"metallb-operator-webhook-server-6bc96cc64c-zfqmv\" (UID: \"85db5798-c470-4ec9-91c3-c9bdd9db327d\") " pod="metallb-system/metallb-operator-webhook-server-6bc96cc64c-zfqmv" Nov 23 06:55:06 crc kubenswrapper[4559]: I1123 06:55:06.032876 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6bc96cc64c-zfqmv" Nov 23 06:55:06 crc kubenswrapper[4559]: I1123 06:55:06.068213 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5c8ccb79b7-ngqhc"] Nov 23 06:55:06 crc kubenswrapper[4559]: W1123 06:55:06.078533 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode57be21d_6409_4c23_b745_ec26bdf7e98d.slice/crio-9aa94f272174dc4ed91da370ad2184ba916f48433818aec9ca2846c0a1df569f WatchSource:0}: Error finding container 9aa94f272174dc4ed91da370ad2184ba916f48433818aec9ca2846c0a1df569f: Status 404 returned error can't find the container with id 9aa94f272174dc4ed91da370ad2184ba916f48433818aec9ca2846c0a1df569f Nov 23 06:55:06 crc kubenswrapper[4559]: I1123 06:55:06.349673 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5c8ccb79b7-ngqhc" event={"ID":"e57be21d-6409-4c23-b745-ec26bdf7e98d","Type":"ContainerStarted","Data":"9aa94f272174dc4ed91da370ad2184ba916f48433818aec9ca2846c0a1df569f"} Nov 23 06:55:06 crc kubenswrapper[4559]: I1123 06:55:06.418888 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6bc96cc64c-zfqmv"] Nov 23 06:55:06 crc kubenswrapper[4559]: W1123 06:55:06.422845 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85db5798_c470_4ec9_91c3_c9bdd9db327d.slice/crio-1c18698c617f9bc71814d93e30671e61e121b699ccadff91047ef79f30276623 WatchSource:0}: Error finding container 1c18698c617f9bc71814d93e30671e61e121b699ccadff91047ef79f30276623: Status 404 returned error can't find the container with id 1c18698c617f9bc71814d93e30671e61e121b699ccadff91047ef79f30276623 Nov 23 06:55:07 crc kubenswrapper[4559]: I1123 06:55:07.355838 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6bc96cc64c-zfqmv" event={"ID":"85db5798-c470-4ec9-91c3-c9bdd9db327d","Type":"ContainerStarted","Data":"1c18698c617f9bc71814d93e30671e61e121b699ccadff91047ef79f30276623"} Nov 23 06:55:08 crc kubenswrapper[4559]: I1123 06:55:08.369479 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5c8ccb79b7-ngqhc" event={"ID":"e57be21d-6409-4c23-b745-ec26bdf7e98d","Type":"ContainerStarted","Data":"21709ab83d610a5b4fc9cc2e611e1e08510612a7f022b83430d096281935ea92"} Nov 23 06:55:08 crc kubenswrapper[4559]: I1123 06:55:08.369899 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5c8ccb79b7-ngqhc" Nov 23 06:55:08 crc kubenswrapper[4559]: I1123 06:55:08.384387 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5c8ccb79b7-ngqhc" podStartSLOduration=1.258613264 podStartE2EDuration="3.384372092s" podCreationTimestamp="2025-11-23 06:55:05 +0000 UTC" firstStartedPulling="2025-11-23 06:55:06.080389063 +0000 UTC m=+608.102374678" lastFinishedPulling="2025-11-23 06:55:08.206147892 +0000 UTC m=+610.228133506" observedRunningTime="2025-11-23 06:55:08.383401776 +0000 UTC m=+610.405387390" watchObservedRunningTime="2025-11-23 06:55:08.384372092 +0000 UTC m=+610.406357706" Nov 23 06:55:10 crc kubenswrapper[4559]: I1123 06:55:10.383014 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6bc96cc64c-zfqmv" event={"ID":"85db5798-c470-4ec9-91c3-c9bdd9db327d","Type":"ContainerStarted","Data":"3a436d6cded79e7dfe6bf1190cd3f15f2e5023029a98fb4028a9fd97c2959d63"} Nov 23 06:55:10 crc kubenswrapper[4559]: I1123 06:55:10.383348 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-6bc96cc64c-zfqmv" Nov 23 06:55:10 crc kubenswrapper[4559]: I1123 06:55:10.398491 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-6bc96cc64c-zfqmv" podStartSLOduration=1.949383707 podStartE2EDuration="5.398469981s" podCreationTimestamp="2025-11-23 06:55:05 +0000 UTC" firstStartedPulling="2025-11-23 06:55:06.425735712 +0000 UTC m=+608.447721326" lastFinishedPulling="2025-11-23 06:55:09.874821987 +0000 UTC m=+611.896807600" observedRunningTime="2025-11-23 06:55:10.395904434 +0000 UTC m=+612.417890048" watchObservedRunningTime="2025-11-23 06:55:10.398469981 +0000 UTC m=+612.420455595" Nov 23 06:55:26 crc kubenswrapper[4559]: I1123 06:55:26.036783 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-6bc96cc64c-zfqmv" Nov 23 06:55:45 crc kubenswrapper[4559]: I1123 06:55:45.657018 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5c8ccb79b7-ngqhc" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.189243 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-h2xjx"] Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.191247 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-h2xjx" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.191786 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-zwlxs"] Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.192358 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-zwlxs" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.194042 4559 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-xbhmp" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.195810 4559 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.195943 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.195982 4559 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.202766 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-zwlxs"] Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.246889 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-stnln"] Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.254930 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-stnln" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.255810 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/b3f1ab44-d9d7-4005-82b1-249213ea6ee6-reloader\") pod \"frr-k8s-h2xjx\" (UID: \"b3f1ab44-d9d7-4005-82b1-249213ea6ee6\") " pod="metallb-system/frr-k8s-h2xjx" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.255866 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/b3f1ab44-d9d7-4005-82b1-249213ea6ee6-frr-startup\") pod \"frr-k8s-h2xjx\" (UID: \"b3f1ab44-d9d7-4005-82b1-249213ea6ee6\") " pod="metallb-system/frr-k8s-h2xjx" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.255905 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ps7g\" (UniqueName: \"kubernetes.io/projected/5533445d-9129-4b9d-b87a-746df3caefb9-kube-api-access-6ps7g\") pod \"frr-k8s-webhook-server-6998585d5-zwlxs\" (UID: \"5533445d-9129-4b9d-b87a-746df3caefb9\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-zwlxs" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.256171 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5533445d-9129-4b9d-b87a-746df3caefb9-cert\") pod \"frr-k8s-webhook-server-6998585d5-zwlxs\" (UID: \"5533445d-9129-4b9d-b87a-746df3caefb9\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-zwlxs" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.256213 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/b3f1ab44-d9d7-4005-82b1-249213ea6ee6-frr-sockets\") pod \"frr-k8s-h2xjx\" (UID: \"b3f1ab44-d9d7-4005-82b1-249213ea6ee6\") " pod="metallb-system/frr-k8s-h2xjx" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.256232 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/b3f1ab44-d9d7-4005-82b1-249213ea6ee6-frr-conf\") pod \"frr-k8s-h2xjx\" (UID: \"b3f1ab44-d9d7-4005-82b1-249213ea6ee6\") " pod="metallb-system/frr-k8s-h2xjx" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.256277 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/b3f1ab44-d9d7-4005-82b1-249213ea6ee6-metrics\") pod \"frr-k8s-h2xjx\" (UID: \"b3f1ab44-d9d7-4005-82b1-249213ea6ee6\") " pod="metallb-system/frr-k8s-h2xjx" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.261065 4559 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.261234 4559 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.261078 4559 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-9j742" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.261533 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.265445 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-tgxdj"] Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.266274 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-tgxdj" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.268585 4559 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.271948 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-tgxdj"] Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.357352 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/24dffaef-1c72-43f9-a5a8-ee0797a63077-cert\") pod \"controller-6c7b4b5f48-tgxdj\" (UID: \"24dffaef-1c72-43f9-a5a8-ee0797a63077\") " pod="metallb-system/controller-6c7b4b5f48-tgxdj" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.357407 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ps7g\" (UniqueName: \"kubernetes.io/projected/5533445d-9129-4b9d-b87a-746df3caefb9-kube-api-access-6ps7g\") pod \"frr-k8s-webhook-server-6998585d5-zwlxs\" (UID: \"5533445d-9129-4b9d-b87a-746df3caefb9\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-zwlxs" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.357457 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-955gh\" (UniqueName: \"kubernetes.io/projected/b3f1ab44-d9d7-4005-82b1-249213ea6ee6-kube-api-access-955gh\") pod \"frr-k8s-h2xjx\" (UID: \"b3f1ab44-d9d7-4005-82b1-249213ea6ee6\") " pod="metallb-system/frr-k8s-h2xjx" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.357480 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5533445d-9129-4b9d-b87a-746df3caefb9-cert\") pod \"frr-k8s-webhook-server-6998585d5-zwlxs\" (UID: \"5533445d-9129-4b9d-b87a-746df3caefb9\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-zwlxs" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.357497 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/b3f1ab44-d9d7-4005-82b1-249213ea6ee6-frr-sockets\") pod \"frr-k8s-h2xjx\" (UID: \"b3f1ab44-d9d7-4005-82b1-249213ea6ee6\") " pod="metallb-system/frr-k8s-h2xjx" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.357510 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/b3f1ab44-d9d7-4005-82b1-249213ea6ee6-frr-conf\") pod \"frr-k8s-h2xjx\" (UID: \"b3f1ab44-d9d7-4005-82b1-249213ea6ee6\") " pod="metallb-system/frr-k8s-h2xjx" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.357528 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/fe556537-f24f-4a84-9f79-488938b3f4b5-memberlist\") pod \"speaker-stnln\" (UID: \"fe556537-f24f-4a84-9f79-488938b3f4b5\") " pod="metallb-system/speaker-stnln" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.357545 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbnvq\" (UniqueName: \"kubernetes.io/projected/24dffaef-1c72-43f9-a5a8-ee0797a63077-kube-api-access-sbnvq\") pod \"controller-6c7b4b5f48-tgxdj\" (UID: \"24dffaef-1c72-43f9-a5a8-ee0797a63077\") " pod="metallb-system/controller-6c7b4b5f48-tgxdj" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.357581 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-927ql\" (UniqueName: \"kubernetes.io/projected/fe556537-f24f-4a84-9f79-488938b3f4b5-kube-api-access-927ql\") pod \"speaker-stnln\" (UID: \"fe556537-f24f-4a84-9f79-488938b3f4b5\") " pod="metallb-system/speaker-stnln" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.357599 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/b3f1ab44-d9d7-4005-82b1-249213ea6ee6-metrics\") pod \"frr-k8s-h2xjx\" (UID: \"b3f1ab44-d9d7-4005-82b1-249213ea6ee6\") " pod="metallb-system/frr-k8s-h2xjx" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.357624 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe556537-f24f-4a84-9f79-488938b3f4b5-metrics-certs\") pod \"speaker-stnln\" (UID: \"fe556537-f24f-4a84-9f79-488938b3f4b5\") " pod="metallb-system/speaker-stnln" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.357673 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/fe556537-f24f-4a84-9f79-488938b3f4b5-metallb-excludel2\") pod \"speaker-stnln\" (UID: \"fe556537-f24f-4a84-9f79-488938b3f4b5\") " pod="metallb-system/speaker-stnln" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.357689 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/24dffaef-1c72-43f9-a5a8-ee0797a63077-metrics-certs\") pod \"controller-6c7b4b5f48-tgxdj\" (UID: \"24dffaef-1c72-43f9-a5a8-ee0797a63077\") " pod="metallb-system/controller-6c7b4b5f48-tgxdj" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.357750 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b3f1ab44-d9d7-4005-82b1-249213ea6ee6-metrics-certs\") pod \"frr-k8s-h2xjx\" (UID: \"b3f1ab44-d9d7-4005-82b1-249213ea6ee6\") " pod="metallb-system/frr-k8s-h2xjx" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.357774 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/b3f1ab44-d9d7-4005-82b1-249213ea6ee6-reloader\") pod \"frr-k8s-h2xjx\" (UID: \"b3f1ab44-d9d7-4005-82b1-249213ea6ee6\") " pod="metallb-system/frr-k8s-h2xjx" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.357853 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/b3f1ab44-d9d7-4005-82b1-249213ea6ee6-frr-startup\") pod \"frr-k8s-h2xjx\" (UID: \"b3f1ab44-d9d7-4005-82b1-249213ea6ee6\") " pod="metallb-system/frr-k8s-h2xjx" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.358317 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/b3f1ab44-d9d7-4005-82b1-249213ea6ee6-metrics\") pod \"frr-k8s-h2xjx\" (UID: \"b3f1ab44-d9d7-4005-82b1-249213ea6ee6\") " pod="metallb-system/frr-k8s-h2xjx" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.358405 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/b3f1ab44-d9d7-4005-82b1-249213ea6ee6-frr-conf\") pod \"frr-k8s-h2xjx\" (UID: \"b3f1ab44-d9d7-4005-82b1-249213ea6ee6\") " pod="metallb-system/frr-k8s-h2xjx" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.358452 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/b3f1ab44-d9d7-4005-82b1-249213ea6ee6-frr-sockets\") pod \"frr-k8s-h2xjx\" (UID: \"b3f1ab44-d9d7-4005-82b1-249213ea6ee6\") " pod="metallb-system/frr-k8s-h2xjx" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.358556 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/b3f1ab44-d9d7-4005-82b1-249213ea6ee6-reloader\") pod \"frr-k8s-h2xjx\" (UID: \"b3f1ab44-d9d7-4005-82b1-249213ea6ee6\") " pod="metallb-system/frr-k8s-h2xjx" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.358808 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/b3f1ab44-d9d7-4005-82b1-249213ea6ee6-frr-startup\") pod \"frr-k8s-h2xjx\" (UID: \"b3f1ab44-d9d7-4005-82b1-249213ea6ee6\") " pod="metallb-system/frr-k8s-h2xjx" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.362406 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5533445d-9129-4b9d-b87a-746df3caefb9-cert\") pod \"frr-k8s-webhook-server-6998585d5-zwlxs\" (UID: \"5533445d-9129-4b9d-b87a-746df3caefb9\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-zwlxs" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.371403 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ps7g\" (UniqueName: \"kubernetes.io/projected/5533445d-9129-4b9d-b87a-746df3caefb9-kube-api-access-6ps7g\") pod \"frr-k8s-webhook-server-6998585d5-zwlxs\" (UID: \"5533445d-9129-4b9d-b87a-746df3caefb9\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-zwlxs" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.459953 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-955gh\" (UniqueName: \"kubernetes.io/projected/b3f1ab44-d9d7-4005-82b1-249213ea6ee6-kube-api-access-955gh\") pod \"frr-k8s-h2xjx\" (UID: \"b3f1ab44-d9d7-4005-82b1-249213ea6ee6\") " pod="metallb-system/frr-k8s-h2xjx" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.460009 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/fe556537-f24f-4a84-9f79-488938b3f4b5-memberlist\") pod \"speaker-stnln\" (UID: \"fe556537-f24f-4a84-9f79-488938b3f4b5\") " pod="metallb-system/speaker-stnln" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.460033 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbnvq\" (UniqueName: \"kubernetes.io/projected/24dffaef-1c72-43f9-a5a8-ee0797a63077-kube-api-access-sbnvq\") pod \"controller-6c7b4b5f48-tgxdj\" (UID: \"24dffaef-1c72-43f9-a5a8-ee0797a63077\") " pod="metallb-system/controller-6c7b4b5f48-tgxdj" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.460061 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-927ql\" (UniqueName: \"kubernetes.io/projected/fe556537-f24f-4a84-9f79-488938b3f4b5-kube-api-access-927ql\") pod \"speaker-stnln\" (UID: \"fe556537-f24f-4a84-9f79-488938b3f4b5\") " pod="metallb-system/speaker-stnln" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.460086 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe556537-f24f-4a84-9f79-488938b3f4b5-metrics-certs\") pod \"speaker-stnln\" (UID: \"fe556537-f24f-4a84-9f79-488938b3f4b5\") " pod="metallb-system/speaker-stnln" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.460111 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/fe556537-f24f-4a84-9f79-488938b3f4b5-metallb-excludel2\") pod \"speaker-stnln\" (UID: \"fe556537-f24f-4a84-9f79-488938b3f4b5\") " pod="metallb-system/speaker-stnln" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.460124 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/24dffaef-1c72-43f9-a5a8-ee0797a63077-metrics-certs\") pod \"controller-6c7b4b5f48-tgxdj\" (UID: \"24dffaef-1c72-43f9-a5a8-ee0797a63077\") " pod="metallb-system/controller-6c7b4b5f48-tgxdj" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.460164 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b3f1ab44-d9d7-4005-82b1-249213ea6ee6-metrics-certs\") pod \"frr-k8s-h2xjx\" (UID: \"b3f1ab44-d9d7-4005-82b1-249213ea6ee6\") " pod="metallb-system/frr-k8s-h2xjx" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.460207 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/24dffaef-1c72-43f9-a5a8-ee0797a63077-cert\") pod \"controller-6c7b4b5f48-tgxdj\" (UID: \"24dffaef-1c72-43f9-a5a8-ee0797a63077\") " pod="metallb-system/controller-6c7b4b5f48-tgxdj" Nov 23 06:55:46 crc kubenswrapper[4559]: E1123 06:55:46.460899 4559 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 23 06:55:46 crc kubenswrapper[4559]: E1123 06:55:46.460958 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe556537-f24f-4a84-9f79-488938b3f4b5-memberlist podName:fe556537-f24f-4a84-9f79-488938b3f4b5 nodeName:}" failed. No retries permitted until 2025-11-23 06:55:46.960942874 +0000 UTC m=+648.982928498 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/fe556537-f24f-4a84-9f79-488938b3f4b5-memberlist") pod "speaker-stnln" (UID: "fe556537-f24f-4a84-9f79-488938b3f4b5") : secret "metallb-memberlist" not found Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.461529 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/fe556537-f24f-4a84-9f79-488938b3f4b5-metallb-excludel2\") pod \"speaker-stnln\" (UID: \"fe556537-f24f-4a84-9f79-488938b3f4b5\") " pod="metallb-system/speaker-stnln" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.462878 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe556537-f24f-4a84-9f79-488938b3f4b5-metrics-certs\") pod \"speaker-stnln\" (UID: \"fe556537-f24f-4a84-9f79-488938b3f4b5\") " pod="metallb-system/speaker-stnln" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.463591 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/24dffaef-1c72-43f9-a5a8-ee0797a63077-cert\") pod \"controller-6c7b4b5f48-tgxdj\" (UID: \"24dffaef-1c72-43f9-a5a8-ee0797a63077\") " pod="metallb-system/controller-6c7b4b5f48-tgxdj" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.465329 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/24dffaef-1c72-43f9-a5a8-ee0797a63077-metrics-certs\") pod \"controller-6c7b4b5f48-tgxdj\" (UID: \"24dffaef-1c72-43f9-a5a8-ee0797a63077\") " pod="metallb-system/controller-6c7b4b5f48-tgxdj" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.465393 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b3f1ab44-d9d7-4005-82b1-249213ea6ee6-metrics-certs\") pod \"frr-k8s-h2xjx\" (UID: \"b3f1ab44-d9d7-4005-82b1-249213ea6ee6\") " pod="metallb-system/frr-k8s-h2xjx" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.472595 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbnvq\" (UniqueName: \"kubernetes.io/projected/24dffaef-1c72-43f9-a5a8-ee0797a63077-kube-api-access-sbnvq\") pod \"controller-6c7b4b5f48-tgxdj\" (UID: \"24dffaef-1c72-43f9-a5a8-ee0797a63077\") " pod="metallb-system/controller-6c7b4b5f48-tgxdj" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.472997 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-927ql\" (UniqueName: \"kubernetes.io/projected/fe556537-f24f-4a84-9f79-488938b3f4b5-kube-api-access-927ql\") pod \"speaker-stnln\" (UID: \"fe556537-f24f-4a84-9f79-488938b3f4b5\") " pod="metallb-system/speaker-stnln" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.473305 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-955gh\" (UniqueName: \"kubernetes.io/projected/b3f1ab44-d9d7-4005-82b1-249213ea6ee6-kube-api-access-955gh\") pod \"frr-k8s-h2xjx\" (UID: \"b3f1ab44-d9d7-4005-82b1-249213ea6ee6\") " pod="metallb-system/frr-k8s-h2xjx" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.505469 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-h2xjx" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.512680 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-zwlxs" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.585833 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-tgxdj" Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.852548 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-zwlxs"] Nov 23 06:55:46 crc kubenswrapper[4559]: W1123 06:55:46.856089 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5533445d_9129_4b9d_b87a_746df3caefb9.slice/crio-f444e0b3fc0a46873729cb56ef701dbc097a2030b9b428f7dbbd20bc5dafa7f7 WatchSource:0}: Error finding container f444e0b3fc0a46873729cb56ef701dbc097a2030b9b428f7dbbd20bc5dafa7f7: Status 404 returned error can't find the container with id f444e0b3fc0a46873729cb56ef701dbc097a2030b9b428f7dbbd20bc5dafa7f7 Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.926795 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-tgxdj"] Nov 23 06:55:46 crc kubenswrapper[4559]: W1123 06:55:46.933433 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24dffaef_1c72_43f9_a5a8_ee0797a63077.slice/crio-58fa2e19fb3478bb0c4d7939c22c0bb2032b2ab5582a8842e991f619fb0efdf1 WatchSource:0}: Error finding container 58fa2e19fb3478bb0c4d7939c22c0bb2032b2ab5582a8842e991f619fb0efdf1: Status 404 returned error can't find the container with id 58fa2e19fb3478bb0c4d7939c22c0bb2032b2ab5582a8842e991f619fb0efdf1 Nov 23 06:55:46 crc kubenswrapper[4559]: I1123 06:55:46.965957 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/fe556537-f24f-4a84-9f79-488938b3f4b5-memberlist\") pod \"speaker-stnln\" (UID: \"fe556537-f24f-4a84-9f79-488938b3f4b5\") " pod="metallb-system/speaker-stnln" Nov 23 06:55:46 crc kubenswrapper[4559]: E1123 06:55:46.966085 4559 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 23 06:55:46 crc kubenswrapper[4559]: E1123 06:55:46.966127 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe556537-f24f-4a84-9f79-488938b3f4b5-memberlist podName:fe556537-f24f-4a84-9f79-488938b3f4b5 nodeName:}" failed. No retries permitted until 2025-11-23 06:55:47.96611505 +0000 UTC m=+649.988100664 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/fe556537-f24f-4a84-9f79-488938b3f4b5-memberlist") pod "speaker-stnln" (UID: "fe556537-f24f-4a84-9f79-488938b3f4b5") : secret "metallb-memberlist" not found Nov 23 06:55:47 crc kubenswrapper[4559]: I1123 06:55:47.552349 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-tgxdj" event={"ID":"24dffaef-1c72-43f9-a5a8-ee0797a63077","Type":"ContainerStarted","Data":"20d258ce2eda446ec3786d519db1f6873546c4720818ff7323c2e1eaf9fc5483"} Nov 23 06:55:47 crc kubenswrapper[4559]: I1123 06:55:47.552577 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-tgxdj" event={"ID":"24dffaef-1c72-43f9-a5a8-ee0797a63077","Type":"ContainerStarted","Data":"5e5be0404616244f6a9c7fd8b82e7cf6b6c3cac1179b31717eba645cd3faece0"} Nov 23 06:55:47 crc kubenswrapper[4559]: I1123 06:55:47.552592 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-tgxdj" event={"ID":"24dffaef-1c72-43f9-a5a8-ee0797a63077","Type":"ContainerStarted","Data":"58fa2e19fb3478bb0c4d7939c22c0bb2032b2ab5582a8842e991f619fb0efdf1"} Nov 23 06:55:47 crc kubenswrapper[4559]: I1123 06:55:47.552606 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-tgxdj" Nov 23 06:55:47 crc kubenswrapper[4559]: I1123 06:55:47.553318 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-zwlxs" event={"ID":"5533445d-9129-4b9d-b87a-746df3caefb9","Type":"ContainerStarted","Data":"f444e0b3fc0a46873729cb56ef701dbc097a2030b9b428f7dbbd20bc5dafa7f7"} Nov 23 06:55:47 crc kubenswrapper[4559]: I1123 06:55:47.554018 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-h2xjx" event={"ID":"b3f1ab44-d9d7-4005-82b1-249213ea6ee6","Type":"ContainerStarted","Data":"2fdda986a93862a8c41be89843a5dd0ea727ac7752f486adee3324769a423ca9"} Nov 23 06:55:47 crc kubenswrapper[4559]: I1123 06:55:47.564031 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-tgxdj" podStartSLOduration=1.5640198330000001 podStartE2EDuration="1.564019833s" podCreationTimestamp="2025-11-23 06:55:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:55:47.563730168 +0000 UTC m=+649.585715782" watchObservedRunningTime="2025-11-23 06:55:47.564019833 +0000 UTC m=+649.586005448" Nov 23 06:55:47 crc kubenswrapper[4559]: I1123 06:55:47.977200 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/fe556537-f24f-4a84-9f79-488938b3f4b5-memberlist\") pod \"speaker-stnln\" (UID: \"fe556537-f24f-4a84-9f79-488938b3f4b5\") " pod="metallb-system/speaker-stnln" Nov 23 06:55:47 crc kubenswrapper[4559]: I1123 06:55:47.983179 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/fe556537-f24f-4a84-9f79-488938b3f4b5-memberlist\") pod \"speaker-stnln\" (UID: \"fe556537-f24f-4a84-9f79-488938b3f4b5\") " pod="metallb-system/speaker-stnln" Nov 23 06:55:48 crc kubenswrapper[4559]: I1123 06:55:48.074124 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-stnln" Nov 23 06:55:48 crc kubenswrapper[4559]: W1123 06:55:48.094142 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfe556537_f24f_4a84_9f79_488938b3f4b5.slice/crio-3b76bf08784e2c649a81928fd029967adfaf921bee0b5d7e9b9766314110a349 WatchSource:0}: Error finding container 3b76bf08784e2c649a81928fd029967adfaf921bee0b5d7e9b9766314110a349: Status 404 returned error can't find the container with id 3b76bf08784e2c649a81928fd029967adfaf921bee0b5d7e9b9766314110a349 Nov 23 06:55:48 crc kubenswrapper[4559]: I1123 06:55:48.568964 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-stnln" event={"ID":"fe556537-f24f-4a84-9f79-488938b3f4b5","Type":"ContainerStarted","Data":"a9c0ac35a19846a8758f137cd1a17cffa7c7d632837f5bb2d8f693ff5c17979c"} Nov 23 06:55:48 crc kubenswrapper[4559]: I1123 06:55:48.569003 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-stnln" event={"ID":"fe556537-f24f-4a84-9f79-488938b3f4b5","Type":"ContainerStarted","Data":"fcf24111e1dd3f2d5357368a6977982fbfb9d6c89b7359631a0c32ec25e7b6c6"} Nov 23 06:55:48 crc kubenswrapper[4559]: I1123 06:55:48.569016 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-stnln" event={"ID":"fe556537-f24f-4a84-9f79-488938b3f4b5","Type":"ContainerStarted","Data":"3b76bf08784e2c649a81928fd029967adfaf921bee0b5d7e9b9766314110a349"} Nov 23 06:55:48 crc kubenswrapper[4559]: I1123 06:55:48.569197 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-stnln" Nov 23 06:55:48 crc kubenswrapper[4559]: I1123 06:55:48.587776 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-stnln" podStartSLOduration=2.58776348 podStartE2EDuration="2.58776348s" podCreationTimestamp="2025-11-23 06:55:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:55:48.586173197 +0000 UTC m=+650.608158831" watchObservedRunningTime="2025-11-23 06:55:48.58776348 +0000 UTC m=+650.609749094" Nov 23 06:55:52 crc kubenswrapper[4559]: I1123 06:55:52.591617 4559 generic.go:334] "Generic (PLEG): container finished" podID="b3f1ab44-d9d7-4005-82b1-249213ea6ee6" containerID="896ad0220116041df7e56d3eaa6461d045719080d4c4c1421e1f37473fc96629" exitCode=0 Nov 23 06:55:52 crc kubenswrapper[4559]: I1123 06:55:52.591677 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-h2xjx" event={"ID":"b3f1ab44-d9d7-4005-82b1-249213ea6ee6","Type":"ContainerDied","Data":"896ad0220116041df7e56d3eaa6461d045719080d4c4c1421e1f37473fc96629"} Nov 23 06:55:52 crc kubenswrapper[4559]: I1123 06:55:52.593422 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-zwlxs" event={"ID":"5533445d-9129-4b9d-b87a-746df3caefb9","Type":"ContainerStarted","Data":"235cc448dc61ae309c26772f8ed579e6ed39a3e0cb0b812b207cc1c4c8a09a1a"} Nov 23 06:55:52 crc kubenswrapper[4559]: I1123 06:55:52.593551 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-zwlxs" Nov 23 06:55:52 crc kubenswrapper[4559]: I1123 06:55:52.616825 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-zwlxs" podStartSLOduration=1.2980714070000001 podStartE2EDuration="6.616811354s" podCreationTimestamp="2025-11-23 06:55:46 +0000 UTC" firstStartedPulling="2025-11-23 06:55:46.857720788 +0000 UTC m=+648.879706402" lastFinishedPulling="2025-11-23 06:55:52.176460735 +0000 UTC m=+654.198446349" observedRunningTime="2025-11-23 06:55:52.615706125 +0000 UTC m=+654.637691739" watchObservedRunningTime="2025-11-23 06:55:52.616811354 +0000 UTC m=+654.638796968" Nov 23 06:55:53 crc kubenswrapper[4559]: I1123 06:55:53.599633 4559 generic.go:334] "Generic (PLEG): container finished" podID="b3f1ab44-d9d7-4005-82b1-249213ea6ee6" containerID="2f916b4ea3b3de0650d71d13cc64fc13de0d04516991d4e7dc1f06e8d82e1fe6" exitCode=0 Nov 23 06:55:53 crc kubenswrapper[4559]: I1123 06:55:53.599694 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-h2xjx" event={"ID":"b3f1ab44-d9d7-4005-82b1-249213ea6ee6","Type":"ContainerDied","Data":"2f916b4ea3b3de0650d71d13cc64fc13de0d04516991d4e7dc1f06e8d82e1fe6"} Nov 23 06:55:54 crc kubenswrapper[4559]: I1123 06:55:54.607000 4559 generic.go:334] "Generic (PLEG): container finished" podID="b3f1ab44-d9d7-4005-82b1-249213ea6ee6" containerID="2b1e0649d4aee721b71cd88acd768036bf259a4c6a6a8f405e2215b79ba3b83e" exitCode=0 Nov 23 06:55:54 crc kubenswrapper[4559]: I1123 06:55:54.607061 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-h2xjx" event={"ID":"b3f1ab44-d9d7-4005-82b1-249213ea6ee6","Type":"ContainerDied","Data":"2b1e0649d4aee721b71cd88acd768036bf259a4c6a6a8f405e2215b79ba3b83e"} Nov 23 06:55:55 crc kubenswrapper[4559]: I1123 06:55:55.617454 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-h2xjx" event={"ID":"b3f1ab44-d9d7-4005-82b1-249213ea6ee6","Type":"ContainerStarted","Data":"20672652e0309773c699757c5210d998f80b7ab9dc8fa0981646e72c412f1df0"} Nov 23 06:55:55 crc kubenswrapper[4559]: I1123 06:55:55.617830 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-h2xjx" event={"ID":"b3f1ab44-d9d7-4005-82b1-249213ea6ee6","Type":"ContainerStarted","Data":"163b66d239bf18a30fbe88d613f1a2ee1fa431a05e7d337d9960a9ab6665544c"} Nov 23 06:55:55 crc kubenswrapper[4559]: I1123 06:55:55.617842 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-h2xjx" event={"ID":"b3f1ab44-d9d7-4005-82b1-249213ea6ee6","Type":"ContainerStarted","Data":"6703fa456014eb1cd1b7159e94ee2287f3c794768b20fff66b945c14772ca250"} Nov 23 06:55:55 crc kubenswrapper[4559]: I1123 06:55:55.617859 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-h2xjx" Nov 23 06:55:55 crc kubenswrapper[4559]: I1123 06:55:55.617870 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-h2xjx" event={"ID":"b3f1ab44-d9d7-4005-82b1-249213ea6ee6","Type":"ContainerStarted","Data":"7d646cebdd042685d543d051b1692754fb3ae1a2917fb30967e66d41b9b70d40"} Nov 23 06:55:55 crc kubenswrapper[4559]: I1123 06:55:55.617879 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-h2xjx" event={"ID":"b3f1ab44-d9d7-4005-82b1-249213ea6ee6","Type":"ContainerStarted","Data":"39670d96a9bad5b77407fcf42e92a87e19c9a670c3986ffd06e25a4745e57bb0"} Nov 23 06:55:55 crc kubenswrapper[4559]: I1123 06:55:55.617888 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-h2xjx" event={"ID":"b3f1ab44-d9d7-4005-82b1-249213ea6ee6","Type":"ContainerStarted","Data":"8c0373853859abb8d7fd7b2ddd0847104d76b0aae2cc13207351b7132044b9f1"} Nov 23 06:55:55 crc kubenswrapper[4559]: I1123 06:55:55.635986 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-h2xjx" podStartSLOduration=4.066999775 podStartE2EDuration="9.635963321s" podCreationTimestamp="2025-11-23 06:55:46 +0000 UTC" firstStartedPulling="2025-11-23 06:55:46.594092781 +0000 UTC m=+648.616078395" lastFinishedPulling="2025-11-23 06:55:52.163056327 +0000 UTC m=+654.185041941" observedRunningTime="2025-11-23 06:55:55.633932842 +0000 UTC m=+657.655918455" watchObservedRunningTime="2025-11-23 06:55:55.635963321 +0000 UTC m=+657.657948935" Nov 23 06:55:56 crc kubenswrapper[4559]: I1123 06:55:56.506457 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-h2xjx" Nov 23 06:55:56 crc kubenswrapper[4559]: I1123 06:55:56.534414 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-h2xjx" Nov 23 06:55:58 crc kubenswrapper[4559]: I1123 06:55:58.077513 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-stnln" Nov 23 06:56:00 crc kubenswrapper[4559]: I1123 06:56:00.200098 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-ksvdl"] Nov 23 06:56:00 crc kubenswrapper[4559]: I1123 06:56:00.200789 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-ksvdl" Nov 23 06:56:00 crc kubenswrapper[4559]: I1123 06:56:00.203273 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 23 06:56:00 crc kubenswrapper[4559]: I1123 06:56:00.203630 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 23 06:56:00 crc kubenswrapper[4559]: I1123 06:56:00.204464 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-49tgz" Nov 23 06:56:00 crc kubenswrapper[4559]: I1123 06:56:00.211155 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-ksvdl"] Nov 23 06:56:00 crc kubenswrapper[4559]: I1123 06:56:00.249753 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4l4d\" (UniqueName: \"kubernetes.io/projected/21e35ffd-4620-4ef0-9192-c3242022f982-kube-api-access-s4l4d\") pod \"openstack-operator-index-ksvdl\" (UID: \"21e35ffd-4620-4ef0-9192-c3242022f982\") " pod="openstack-operators/openstack-operator-index-ksvdl" Nov 23 06:56:00 crc kubenswrapper[4559]: I1123 06:56:00.350627 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4l4d\" (UniqueName: \"kubernetes.io/projected/21e35ffd-4620-4ef0-9192-c3242022f982-kube-api-access-s4l4d\") pod \"openstack-operator-index-ksvdl\" (UID: \"21e35ffd-4620-4ef0-9192-c3242022f982\") " pod="openstack-operators/openstack-operator-index-ksvdl" Nov 23 06:56:00 crc kubenswrapper[4559]: I1123 06:56:00.365889 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4l4d\" (UniqueName: \"kubernetes.io/projected/21e35ffd-4620-4ef0-9192-c3242022f982-kube-api-access-s4l4d\") pod \"openstack-operator-index-ksvdl\" (UID: \"21e35ffd-4620-4ef0-9192-c3242022f982\") " pod="openstack-operators/openstack-operator-index-ksvdl" Nov 23 06:56:00 crc kubenswrapper[4559]: I1123 06:56:00.514283 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-ksvdl" Nov 23 06:56:00 crc kubenswrapper[4559]: I1123 06:56:00.864929 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-ksvdl"] Nov 23 06:56:00 crc kubenswrapper[4559]: W1123 06:56:00.871590 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21e35ffd_4620_4ef0_9192_c3242022f982.slice/crio-a60d82d039f8b258972f220a1cdc449662b9f9caf31e1ad2f88baf7bf77fcf88 WatchSource:0}: Error finding container a60d82d039f8b258972f220a1cdc449662b9f9caf31e1ad2f88baf7bf77fcf88: Status 404 returned error can't find the container with id a60d82d039f8b258972f220a1cdc449662b9f9caf31e1ad2f88baf7bf77fcf88 Nov 23 06:56:01 crc kubenswrapper[4559]: I1123 06:56:01.649325 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-ksvdl" event={"ID":"21e35ffd-4620-4ef0-9192-c3242022f982","Type":"ContainerStarted","Data":"bb38a4477d4a14430540486943f93aa26426203c95bdaaac946f53295c52620f"} Nov 23 06:56:01 crc kubenswrapper[4559]: I1123 06:56:01.649668 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-ksvdl" event={"ID":"21e35ffd-4620-4ef0-9192-c3242022f982","Type":"ContainerStarted","Data":"a60d82d039f8b258972f220a1cdc449662b9f9caf31e1ad2f88baf7bf77fcf88"} Nov 23 06:56:01 crc kubenswrapper[4559]: I1123 06:56:01.662533 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-ksvdl" podStartSLOduration=1.000514575 podStartE2EDuration="1.662512994s" podCreationTimestamp="2025-11-23 06:56:00 +0000 UTC" firstStartedPulling="2025-11-23 06:56:00.873246576 +0000 UTC m=+662.895232190" lastFinishedPulling="2025-11-23 06:56:01.535244995 +0000 UTC m=+663.557230609" observedRunningTime="2025-11-23 06:56:01.659313387 +0000 UTC m=+663.681299000" watchObservedRunningTime="2025-11-23 06:56:01.662512994 +0000 UTC m=+663.684498609" Nov 23 06:56:03 crc kubenswrapper[4559]: I1123 06:56:03.585366 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-ksvdl"] Nov 23 06:56:03 crc kubenswrapper[4559]: I1123 06:56:03.658461 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-ksvdl" podUID="21e35ffd-4620-4ef0-9192-c3242022f982" containerName="registry-server" containerID="cri-o://bb38a4477d4a14430540486943f93aa26426203c95bdaaac946f53295c52620f" gracePeriod=2 Nov 23 06:56:03 crc kubenswrapper[4559]: I1123 06:56:03.957384 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-ksvdl" Nov 23 06:56:03 crc kubenswrapper[4559]: I1123 06:56:03.996954 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4l4d\" (UniqueName: \"kubernetes.io/projected/21e35ffd-4620-4ef0-9192-c3242022f982-kube-api-access-s4l4d\") pod \"21e35ffd-4620-4ef0-9192-c3242022f982\" (UID: \"21e35ffd-4620-4ef0-9192-c3242022f982\") " Nov 23 06:56:04 crc kubenswrapper[4559]: I1123 06:56:04.002490 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21e35ffd-4620-4ef0-9192-c3242022f982-kube-api-access-s4l4d" (OuterVolumeSpecName: "kube-api-access-s4l4d") pod "21e35ffd-4620-4ef0-9192-c3242022f982" (UID: "21e35ffd-4620-4ef0-9192-c3242022f982"). InnerVolumeSpecName "kube-api-access-s4l4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:56:04 crc kubenswrapper[4559]: I1123 06:56:04.098999 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4l4d\" (UniqueName: \"kubernetes.io/projected/21e35ffd-4620-4ef0-9192-c3242022f982-kube-api-access-s4l4d\") on node \"crc\" DevicePath \"\"" Nov 23 06:56:04 crc kubenswrapper[4559]: I1123 06:56:04.192496 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-clkrn"] Nov 23 06:56:04 crc kubenswrapper[4559]: E1123 06:56:04.192763 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21e35ffd-4620-4ef0-9192-c3242022f982" containerName="registry-server" Nov 23 06:56:04 crc kubenswrapper[4559]: I1123 06:56:04.192795 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="21e35ffd-4620-4ef0-9192-c3242022f982" containerName="registry-server" Nov 23 06:56:04 crc kubenswrapper[4559]: I1123 06:56:04.192989 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="21e35ffd-4620-4ef0-9192-c3242022f982" containerName="registry-server" Nov 23 06:56:04 crc kubenswrapper[4559]: I1123 06:56:04.193342 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-clkrn" Nov 23 06:56:04 crc kubenswrapper[4559]: I1123 06:56:04.197246 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-clkrn"] Nov 23 06:56:04 crc kubenswrapper[4559]: I1123 06:56:04.302370 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxbdw\" (UniqueName: \"kubernetes.io/projected/7d86f9e5-a7b8-4807-bdeb-149b868a3c7a-kube-api-access-zxbdw\") pod \"openstack-operator-index-clkrn\" (UID: \"7d86f9e5-a7b8-4807-bdeb-149b868a3c7a\") " pod="openstack-operators/openstack-operator-index-clkrn" Nov 23 06:56:04 crc kubenswrapper[4559]: I1123 06:56:04.404215 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxbdw\" (UniqueName: \"kubernetes.io/projected/7d86f9e5-a7b8-4807-bdeb-149b868a3c7a-kube-api-access-zxbdw\") pod \"openstack-operator-index-clkrn\" (UID: \"7d86f9e5-a7b8-4807-bdeb-149b868a3c7a\") " pod="openstack-operators/openstack-operator-index-clkrn" Nov 23 06:56:04 crc kubenswrapper[4559]: I1123 06:56:04.419160 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxbdw\" (UniqueName: \"kubernetes.io/projected/7d86f9e5-a7b8-4807-bdeb-149b868a3c7a-kube-api-access-zxbdw\") pod \"openstack-operator-index-clkrn\" (UID: \"7d86f9e5-a7b8-4807-bdeb-149b868a3c7a\") " pod="openstack-operators/openstack-operator-index-clkrn" Nov 23 06:56:04 crc kubenswrapper[4559]: I1123 06:56:04.507314 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-clkrn" Nov 23 06:56:04 crc kubenswrapper[4559]: I1123 06:56:04.664580 4559 generic.go:334] "Generic (PLEG): container finished" podID="21e35ffd-4620-4ef0-9192-c3242022f982" containerID="bb38a4477d4a14430540486943f93aa26426203c95bdaaac946f53295c52620f" exitCode=0 Nov 23 06:56:04 crc kubenswrapper[4559]: I1123 06:56:04.664669 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-ksvdl" Nov 23 06:56:04 crc kubenswrapper[4559]: I1123 06:56:04.664695 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-ksvdl" event={"ID":"21e35ffd-4620-4ef0-9192-c3242022f982","Type":"ContainerDied","Data":"bb38a4477d4a14430540486943f93aa26426203c95bdaaac946f53295c52620f"} Nov 23 06:56:04 crc kubenswrapper[4559]: I1123 06:56:04.665003 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-ksvdl" event={"ID":"21e35ffd-4620-4ef0-9192-c3242022f982","Type":"ContainerDied","Data":"a60d82d039f8b258972f220a1cdc449662b9f9caf31e1ad2f88baf7bf77fcf88"} Nov 23 06:56:04 crc kubenswrapper[4559]: I1123 06:56:04.665044 4559 scope.go:117] "RemoveContainer" containerID="bb38a4477d4a14430540486943f93aa26426203c95bdaaac946f53295c52620f" Nov 23 06:56:04 crc kubenswrapper[4559]: I1123 06:56:04.680383 4559 scope.go:117] "RemoveContainer" containerID="bb38a4477d4a14430540486943f93aa26426203c95bdaaac946f53295c52620f" Nov 23 06:56:04 crc kubenswrapper[4559]: E1123 06:56:04.680870 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb38a4477d4a14430540486943f93aa26426203c95bdaaac946f53295c52620f\": container with ID starting with bb38a4477d4a14430540486943f93aa26426203c95bdaaac946f53295c52620f not found: ID does not exist" containerID="bb38a4477d4a14430540486943f93aa26426203c95bdaaac946f53295c52620f" Nov 23 06:56:04 crc kubenswrapper[4559]: I1123 06:56:04.680928 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb38a4477d4a14430540486943f93aa26426203c95bdaaac946f53295c52620f"} err="failed to get container status \"bb38a4477d4a14430540486943f93aa26426203c95bdaaac946f53295c52620f\": rpc error: code = NotFound desc = could not find container \"bb38a4477d4a14430540486943f93aa26426203c95bdaaac946f53295c52620f\": container with ID starting with bb38a4477d4a14430540486943f93aa26426203c95bdaaac946f53295c52620f not found: ID does not exist" Nov 23 06:56:04 crc kubenswrapper[4559]: I1123 06:56:04.683844 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-ksvdl"] Nov 23 06:56:04 crc kubenswrapper[4559]: I1123 06:56:04.686952 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-ksvdl"] Nov 23 06:56:04 crc kubenswrapper[4559]: I1123 06:56:04.864718 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-clkrn"] Nov 23 06:56:04 crc kubenswrapper[4559]: W1123 06:56:04.868552 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d86f9e5_a7b8_4807_bdeb_149b868a3c7a.slice/crio-188991781b26ec1f5125989c3056ea08597e44f21ce429c56864acbb0b3309f2 WatchSource:0}: Error finding container 188991781b26ec1f5125989c3056ea08597e44f21ce429c56864acbb0b3309f2: Status 404 returned error can't find the container with id 188991781b26ec1f5125989c3056ea08597e44f21ce429c56864acbb0b3309f2 Nov 23 06:56:05 crc kubenswrapper[4559]: I1123 06:56:05.673018 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-clkrn" event={"ID":"7d86f9e5-a7b8-4807-bdeb-149b868a3c7a","Type":"ContainerStarted","Data":"aa70b3774873ff2f4265148583746deb32a85b684c48a133bf2866148d3a39ac"} Nov 23 06:56:05 crc kubenswrapper[4559]: I1123 06:56:05.673067 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-clkrn" event={"ID":"7d86f9e5-a7b8-4807-bdeb-149b868a3c7a","Type":"ContainerStarted","Data":"188991781b26ec1f5125989c3056ea08597e44f21ce429c56864acbb0b3309f2"} Nov 23 06:56:05 crc kubenswrapper[4559]: I1123 06:56:05.690876 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-clkrn" podStartSLOduration=1.188904585 podStartE2EDuration="1.69085596s" podCreationTimestamp="2025-11-23 06:56:04 +0000 UTC" firstStartedPulling="2025-11-23 06:56:04.872403456 +0000 UTC m=+666.894389069" lastFinishedPulling="2025-11-23 06:56:05.374354839 +0000 UTC m=+667.396340444" observedRunningTime="2025-11-23 06:56:05.686406489 +0000 UTC m=+667.708392103" watchObservedRunningTime="2025-11-23 06:56:05.69085596 +0000 UTC m=+667.712841574" Nov 23 06:56:06 crc kubenswrapper[4559]: I1123 06:56:06.285570 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21e35ffd-4620-4ef0-9192-c3242022f982" path="/var/lib/kubelet/pods/21e35ffd-4620-4ef0-9192-c3242022f982/volumes" Nov 23 06:56:06 crc kubenswrapper[4559]: I1123 06:56:06.509458 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-h2xjx" Nov 23 06:56:06 crc kubenswrapper[4559]: I1123 06:56:06.520302 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-zwlxs" Nov 23 06:56:06 crc kubenswrapper[4559]: I1123 06:56:06.590007 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-tgxdj" Nov 23 06:56:14 crc kubenswrapper[4559]: I1123 06:56:14.507923 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-clkrn" Nov 23 06:56:14 crc kubenswrapper[4559]: I1123 06:56:14.508489 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-clkrn" Nov 23 06:56:14 crc kubenswrapper[4559]: I1123 06:56:14.536093 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-clkrn" Nov 23 06:56:14 crc kubenswrapper[4559]: I1123 06:56:14.746542 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-clkrn" Nov 23 06:56:16 crc kubenswrapper[4559]: I1123 06:56:16.225394 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt"] Nov 23 06:56:16 crc kubenswrapper[4559]: I1123 06:56:16.226941 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt" Nov 23 06:56:16 crc kubenswrapper[4559]: I1123 06:56:16.228844 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-6hcmf" Nov 23 06:56:16 crc kubenswrapper[4559]: I1123 06:56:16.229735 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aaa08108-3039-4f4c-b01f-93dc7deefaa5-bundle\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt\" (UID: \"aaa08108-3039-4f4c-b01f-93dc7deefaa5\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt" Nov 23 06:56:16 crc kubenswrapper[4559]: I1123 06:56:16.229823 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lx7k4\" (UniqueName: \"kubernetes.io/projected/aaa08108-3039-4f4c-b01f-93dc7deefaa5-kube-api-access-lx7k4\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt\" (UID: \"aaa08108-3039-4f4c-b01f-93dc7deefaa5\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt" Nov 23 06:56:16 crc kubenswrapper[4559]: I1123 06:56:16.229857 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aaa08108-3039-4f4c-b01f-93dc7deefaa5-util\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt\" (UID: \"aaa08108-3039-4f4c-b01f-93dc7deefaa5\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt" Nov 23 06:56:16 crc kubenswrapper[4559]: I1123 06:56:16.238392 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt"] Nov 23 06:56:16 crc kubenswrapper[4559]: I1123 06:56:16.331053 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aaa08108-3039-4f4c-b01f-93dc7deefaa5-util\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt\" (UID: \"aaa08108-3039-4f4c-b01f-93dc7deefaa5\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt" Nov 23 06:56:16 crc kubenswrapper[4559]: I1123 06:56:16.331147 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aaa08108-3039-4f4c-b01f-93dc7deefaa5-bundle\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt\" (UID: \"aaa08108-3039-4f4c-b01f-93dc7deefaa5\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt" Nov 23 06:56:16 crc kubenswrapper[4559]: I1123 06:56:16.331210 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lx7k4\" (UniqueName: \"kubernetes.io/projected/aaa08108-3039-4f4c-b01f-93dc7deefaa5-kube-api-access-lx7k4\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt\" (UID: \"aaa08108-3039-4f4c-b01f-93dc7deefaa5\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt" Nov 23 06:56:16 crc kubenswrapper[4559]: I1123 06:56:16.331564 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aaa08108-3039-4f4c-b01f-93dc7deefaa5-bundle\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt\" (UID: \"aaa08108-3039-4f4c-b01f-93dc7deefaa5\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt" Nov 23 06:56:16 crc kubenswrapper[4559]: I1123 06:56:16.331767 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aaa08108-3039-4f4c-b01f-93dc7deefaa5-util\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt\" (UID: \"aaa08108-3039-4f4c-b01f-93dc7deefaa5\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt" Nov 23 06:56:16 crc kubenswrapper[4559]: I1123 06:56:16.350710 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lx7k4\" (UniqueName: \"kubernetes.io/projected/aaa08108-3039-4f4c-b01f-93dc7deefaa5-kube-api-access-lx7k4\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt\" (UID: \"aaa08108-3039-4f4c-b01f-93dc7deefaa5\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt" Nov 23 06:56:16 crc kubenswrapper[4559]: I1123 06:56:16.543993 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt" Nov 23 06:56:16 crc kubenswrapper[4559]: I1123 06:56:16.913943 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt"] Nov 23 06:56:17 crc kubenswrapper[4559]: I1123 06:56:17.740490 4559 generic.go:334] "Generic (PLEG): container finished" podID="aaa08108-3039-4f4c-b01f-93dc7deefaa5" containerID="7eb7ecf3e773706f09831af6d01a4a4422994cb45a1da3bcfb14930cc74eed87" exitCode=0 Nov 23 06:56:17 crc kubenswrapper[4559]: I1123 06:56:17.740531 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt" event={"ID":"aaa08108-3039-4f4c-b01f-93dc7deefaa5","Type":"ContainerDied","Data":"7eb7ecf3e773706f09831af6d01a4a4422994cb45a1da3bcfb14930cc74eed87"} Nov 23 06:56:17 crc kubenswrapper[4559]: I1123 06:56:17.740811 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt" event={"ID":"aaa08108-3039-4f4c-b01f-93dc7deefaa5","Type":"ContainerStarted","Data":"eecbb0d382cc5ed6732e80c855edb925227ce0aa447f58830ca92236cf585838"} Nov 23 06:56:18 crc kubenswrapper[4559]: I1123 06:56:18.748064 4559 generic.go:334] "Generic (PLEG): container finished" podID="aaa08108-3039-4f4c-b01f-93dc7deefaa5" containerID="01ae4393c649d9b2e07968c6801a0c4d66bd39cb9eca434ce1018ecea50d7a96" exitCode=0 Nov 23 06:56:18 crc kubenswrapper[4559]: I1123 06:56:18.748131 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt" event={"ID":"aaa08108-3039-4f4c-b01f-93dc7deefaa5","Type":"ContainerDied","Data":"01ae4393c649d9b2e07968c6801a0c4d66bd39cb9eca434ce1018ecea50d7a96"} Nov 23 06:56:19 crc kubenswrapper[4559]: I1123 06:56:19.754111 4559 generic.go:334] "Generic (PLEG): container finished" podID="aaa08108-3039-4f4c-b01f-93dc7deefaa5" containerID="9908cdabdd4f29ee4132b2d47c3785a9ba635c4442d630063e513ea09eb9c86f" exitCode=0 Nov 23 06:56:19 crc kubenswrapper[4559]: I1123 06:56:19.754188 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt" event={"ID":"aaa08108-3039-4f4c-b01f-93dc7deefaa5","Type":"ContainerDied","Data":"9908cdabdd4f29ee4132b2d47c3785a9ba635c4442d630063e513ea09eb9c86f"} Nov 23 06:56:20 crc kubenswrapper[4559]: I1123 06:56:20.949120 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt" Nov 23 06:56:21 crc kubenswrapper[4559]: I1123 06:56:21.082715 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aaa08108-3039-4f4c-b01f-93dc7deefaa5-util\") pod \"aaa08108-3039-4f4c-b01f-93dc7deefaa5\" (UID: \"aaa08108-3039-4f4c-b01f-93dc7deefaa5\") " Nov 23 06:56:21 crc kubenswrapper[4559]: I1123 06:56:21.082843 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aaa08108-3039-4f4c-b01f-93dc7deefaa5-bundle\") pod \"aaa08108-3039-4f4c-b01f-93dc7deefaa5\" (UID: \"aaa08108-3039-4f4c-b01f-93dc7deefaa5\") " Nov 23 06:56:21 crc kubenswrapper[4559]: I1123 06:56:21.082930 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lx7k4\" (UniqueName: \"kubernetes.io/projected/aaa08108-3039-4f4c-b01f-93dc7deefaa5-kube-api-access-lx7k4\") pod \"aaa08108-3039-4f4c-b01f-93dc7deefaa5\" (UID: \"aaa08108-3039-4f4c-b01f-93dc7deefaa5\") " Nov 23 06:56:21 crc kubenswrapper[4559]: I1123 06:56:21.083494 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aaa08108-3039-4f4c-b01f-93dc7deefaa5-bundle" (OuterVolumeSpecName: "bundle") pod "aaa08108-3039-4f4c-b01f-93dc7deefaa5" (UID: "aaa08108-3039-4f4c-b01f-93dc7deefaa5"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:56:21 crc kubenswrapper[4559]: I1123 06:56:21.089032 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aaa08108-3039-4f4c-b01f-93dc7deefaa5-kube-api-access-lx7k4" (OuterVolumeSpecName: "kube-api-access-lx7k4") pod "aaa08108-3039-4f4c-b01f-93dc7deefaa5" (UID: "aaa08108-3039-4f4c-b01f-93dc7deefaa5"). InnerVolumeSpecName "kube-api-access-lx7k4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:56:21 crc kubenswrapper[4559]: I1123 06:56:21.092490 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aaa08108-3039-4f4c-b01f-93dc7deefaa5-util" (OuterVolumeSpecName: "util") pod "aaa08108-3039-4f4c-b01f-93dc7deefaa5" (UID: "aaa08108-3039-4f4c-b01f-93dc7deefaa5"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:56:21 crc kubenswrapper[4559]: I1123 06:56:21.183862 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lx7k4\" (UniqueName: \"kubernetes.io/projected/aaa08108-3039-4f4c-b01f-93dc7deefaa5-kube-api-access-lx7k4\") on node \"crc\" DevicePath \"\"" Nov 23 06:56:21 crc kubenswrapper[4559]: I1123 06:56:21.183893 4559 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aaa08108-3039-4f4c-b01f-93dc7deefaa5-util\") on node \"crc\" DevicePath \"\"" Nov 23 06:56:21 crc kubenswrapper[4559]: I1123 06:56:21.183905 4559 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aaa08108-3039-4f4c-b01f-93dc7deefaa5-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:56:21 crc kubenswrapper[4559]: I1123 06:56:21.767604 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt" event={"ID":"aaa08108-3039-4f4c-b01f-93dc7deefaa5","Type":"ContainerDied","Data":"eecbb0d382cc5ed6732e80c855edb925227ce0aa447f58830ca92236cf585838"} Nov 23 06:56:21 crc kubenswrapper[4559]: I1123 06:56:21.767655 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eecbb0d382cc5ed6732e80c855edb925227ce0aa447f58830ca92236cf585838" Nov 23 06:56:21 crc kubenswrapper[4559]: I1123 06:56:21.767687 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt" Nov 23 06:56:26 crc kubenswrapper[4559]: I1123 06:56:26.166994 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 06:56:26 crc kubenswrapper[4559]: I1123 06:56:26.167460 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 06:56:28 crc kubenswrapper[4559]: I1123 06:56:28.686089 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8486c7f98b-tfq8j"] Nov 23 06:56:28 crc kubenswrapper[4559]: E1123 06:56:28.686741 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaa08108-3039-4f4c-b01f-93dc7deefaa5" containerName="extract" Nov 23 06:56:28 crc kubenswrapper[4559]: I1123 06:56:28.686756 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaa08108-3039-4f4c-b01f-93dc7deefaa5" containerName="extract" Nov 23 06:56:28 crc kubenswrapper[4559]: E1123 06:56:28.686777 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaa08108-3039-4f4c-b01f-93dc7deefaa5" containerName="pull" Nov 23 06:56:28 crc kubenswrapper[4559]: I1123 06:56:28.686783 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaa08108-3039-4f4c-b01f-93dc7deefaa5" containerName="pull" Nov 23 06:56:28 crc kubenswrapper[4559]: E1123 06:56:28.686794 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaa08108-3039-4f4c-b01f-93dc7deefaa5" containerName="util" Nov 23 06:56:28 crc kubenswrapper[4559]: I1123 06:56:28.686803 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaa08108-3039-4f4c-b01f-93dc7deefaa5" containerName="util" Nov 23 06:56:28 crc kubenswrapper[4559]: I1123 06:56:28.686915 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaa08108-3039-4f4c-b01f-93dc7deefaa5" containerName="extract" Nov 23 06:56:28 crc kubenswrapper[4559]: I1123 06:56:28.687490 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-tfq8j" Nov 23 06:56:28 crc kubenswrapper[4559]: I1123 06:56:28.692277 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-t7nwb" Nov 23 06:56:28 crc kubenswrapper[4559]: I1123 06:56:28.710718 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8486c7f98b-tfq8j"] Nov 23 06:56:28 crc kubenswrapper[4559]: I1123 06:56:28.814501 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbxqb\" (UniqueName: \"kubernetes.io/projected/febcfb0d-e692-4e93-bbb5-40efe2929587-kube-api-access-pbxqb\") pod \"openstack-operator-controller-operator-8486c7f98b-tfq8j\" (UID: \"febcfb0d-e692-4e93-bbb5-40efe2929587\") " pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-tfq8j" Nov 23 06:56:28 crc kubenswrapper[4559]: I1123 06:56:28.915749 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbxqb\" (UniqueName: \"kubernetes.io/projected/febcfb0d-e692-4e93-bbb5-40efe2929587-kube-api-access-pbxqb\") pod \"openstack-operator-controller-operator-8486c7f98b-tfq8j\" (UID: \"febcfb0d-e692-4e93-bbb5-40efe2929587\") " pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-tfq8j" Nov 23 06:56:28 crc kubenswrapper[4559]: I1123 06:56:28.930935 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbxqb\" (UniqueName: \"kubernetes.io/projected/febcfb0d-e692-4e93-bbb5-40efe2929587-kube-api-access-pbxqb\") pod \"openstack-operator-controller-operator-8486c7f98b-tfq8j\" (UID: \"febcfb0d-e692-4e93-bbb5-40efe2929587\") " pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-tfq8j" Nov 23 06:56:29 crc kubenswrapper[4559]: I1123 06:56:29.002810 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-tfq8j" Nov 23 06:56:29 crc kubenswrapper[4559]: I1123 06:56:29.169499 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8486c7f98b-tfq8j"] Nov 23 06:56:29 crc kubenswrapper[4559]: I1123 06:56:29.803483 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-tfq8j" event={"ID":"febcfb0d-e692-4e93-bbb5-40efe2929587","Type":"ContainerStarted","Data":"8bfc77d7a06f971b648ee2a257b9a2d1e49992954d0b360ba6e7c82c612cde27"} Nov 23 06:56:32 crc kubenswrapper[4559]: I1123 06:56:32.822898 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-tfq8j" event={"ID":"febcfb0d-e692-4e93-bbb5-40efe2929587","Type":"ContainerStarted","Data":"a90f21cd1960663181f19a3469fa7a7a108926bfac595aa684c74b912cccf8cd"} Nov 23 06:56:34 crc kubenswrapper[4559]: I1123 06:56:34.836291 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-tfq8j" event={"ID":"febcfb0d-e692-4e93-bbb5-40efe2929587","Type":"ContainerStarted","Data":"ed409ecd533310fff7e49b7b443606ca17fb92507339d01a092acec07e136104"} Nov 23 06:56:34 crc kubenswrapper[4559]: I1123 06:56:34.836931 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-tfq8j" Nov 23 06:56:34 crc kubenswrapper[4559]: I1123 06:56:34.860166 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-tfq8j" podStartSLOduration=1.8803655639999999 podStartE2EDuration="6.860152701s" podCreationTimestamp="2025-11-23 06:56:28 +0000 UTC" firstStartedPulling="2025-11-23 06:56:29.176159142 +0000 UTC m=+691.198144756" lastFinishedPulling="2025-11-23 06:56:34.155946279 +0000 UTC m=+696.177931893" observedRunningTime="2025-11-23 06:56:34.856604275 +0000 UTC m=+696.878589889" watchObservedRunningTime="2025-11-23 06:56:34.860152701 +0000 UTC m=+696.882138315" Nov 23 06:56:39 crc kubenswrapper[4559]: I1123 06:56:39.005397 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-tfq8j" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.085579 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7768f8c84f-ks2b7"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.087112 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-ks2b7" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.091416 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-p66lk" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.095205 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-d78br"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.096311 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-d78br" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.098743 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7768f8c84f-ks2b7"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.109976 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-dx46d" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.118982 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-d78br"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.122034 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-56dfb6b67f-47bnx"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.122840 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-47bnx" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.133440 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-28v2d" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.166558 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-56dfb6b67f-47bnx"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.173247 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-8667fbf6f6-qfgkh"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.174392 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-qfgkh" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.185145 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-2dlkb" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.186539 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6bfm\" (UniqueName: \"kubernetes.io/projected/090f1d86-b4e0-43ca-9aee-49c771cab8c1-kube-api-access-f6bfm\") pod \"barbican-operator-controller-manager-7768f8c84f-ks2b7\" (UID: \"090f1d86-b4e0-43ca-9aee-49c771cab8c1\") " pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-ks2b7" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.186610 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52f5d\" (UniqueName: \"kubernetes.io/projected/35944557-7948-4d54-8225-788d51eb01d1-kube-api-access-52f5d\") pod \"cinder-operator-controller-manager-6d8fd67bf7-d78br\" (UID: \"35944557-7948-4d54-8225-788d51eb01d1\") " pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-d78br" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.188228 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-8667fbf6f6-qfgkh"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.196752 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-bf4c6585d-9tvg2"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.198216 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-9tvg2" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.199911 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-4db6b" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.208973 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-bf4c6585d-9tvg2"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.213071 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d86b44686-c776j"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.215622 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-c776j" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.217320 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-tfvqv" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.218341 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d86b44686-c776j"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.223204 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-769d9c7585-qq8sm"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.224009 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-qq8sm" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.226998 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.227333 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-gxb87" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.239784 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5c75d7c94b-5tcjp"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.243281 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-5tcjp" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.245311 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-v4wxd" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.247577 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-769d9c7585-qq8sm"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.258683 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5c75d7c94b-5tcjp"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.270356 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7879fb76fd-rxpxw"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.271270 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-rxpxw" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.273073 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-m68xp" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.283708 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7bb88cb858-qgvkm"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.285001 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-qgvkm" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.289083 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bbd5\" (UniqueName: \"kubernetes.io/projected/6f79eaa5-12ee-4c47-8c5f-c0fa9190bc2d-kube-api-access-4bbd5\") pod \"designate-operator-controller-manager-56dfb6b67f-47bnx\" (UID: \"6f79eaa5-12ee-4c47-8c5f-c0fa9190bc2d\") " pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-47bnx" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.289130 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6bfm\" (UniqueName: \"kubernetes.io/projected/090f1d86-b4e0-43ca-9aee-49c771cab8c1-kube-api-access-f6bfm\") pod \"barbican-operator-controller-manager-7768f8c84f-ks2b7\" (UID: \"090f1d86-b4e0-43ca-9aee-49c771cab8c1\") " pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-ks2b7" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.289201 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52f5d\" (UniqueName: \"kubernetes.io/projected/35944557-7948-4d54-8225-788d51eb01d1-kube-api-access-52f5d\") pod \"cinder-operator-controller-manager-6d8fd67bf7-d78br\" (UID: \"35944557-7948-4d54-8225-788d51eb01d1\") " pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-d78br" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.289240 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbsg6\" (UniqueName: \"kubernetes.io/projected/042a59c3-a00e-4934-b16e-c731f0ab903f-kube-api-access-nbsg6\") pod \"glance-operator-controller-manager-8667fbf6f6-qfgkh\" (UID: \"042a59c3-a00e-4934-b16e-c731f0ab903f\") " pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-qfgkh" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.289261 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h42qr\" (UniqueName: \"kubernetes.io/projected/5b70e018-1ab3-4f23-b43f-1f433a096b81-kube-api-access-h42qr\") pod \"heat-operator-controller-manager-bf4c6585d-9tvg2\" (UID: \"5b70e018-1ab3-4f23-b43f-1f433a096b81\") " pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-9tvg2" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.291937 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-rdgxh" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.295773 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7879fb76fd-rxpxw"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.300547 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7bb88cb858-qgvkm"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.315047 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-5rhv2"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.316843 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-5rhv2" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.319293 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6bfm\" (UniqueName: \"kubernetes.io/projected/090f1d86-b4e0-43ca-9aee-49c771cab8c1-kube-api-access-f6bfm\") pod \"barbican-operator-controller-manager-7768f8c84f-ks2b7\" (UID: \"090f1d86-b4e0-43ca-9aee-49c771cab8c1\") " pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-ks2b7" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.321298 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52f5d\" (UniqueName: \"kubernetes.io/projected/35944557-7948-4d54-8225-788d51eb01d1-kube-api-access-52f5d\") pod \"cinder-operator-controller-manager-6d8fd67bf7-d78br\" (UID: \"35944557-7948-4d54-8225-788d51eb01d1\") " pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-d78br" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.323216 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-v2c5m" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.325211 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-66b7d6f598-hgbh8"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.327104 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-hgbh8" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.328532 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-jzglw" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.343692 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-5rhv2"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.343750 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-86d796d84d-6ht8s"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.344770 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-6ht8s" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.346618 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-66b7d6f598-hgbh8"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.346991 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-zpsx4" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.357676 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6fdc856c5d-bg99z"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.362036 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-86d796d84d-6ht8s"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.362128 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-bg99z" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.364411 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-q6g9g" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.370402 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6fdc856c5d-bg99z"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.381057 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-8hrnp"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.382107 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-8hrnp" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.386931 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-5gkmp" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.391482 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-8hrnp"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.392567 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/82279b3c-bca1-4891-b7d9-a367005ad84e-cert\") pod \"infra-operator-controller-manager-769d9c7585-qq8sm\" (UID: \"82279b3c-bca1-4891-b7d9-a367005ad84e\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-qq8sm" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.392599 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-772zl\" (UniqueName: \"kubernetes.io/projected/527a2a9e-741e-4f1b-8546-a852d73a836b-kube-api-access-772zl\") pod \"ironic-operator-controller-manager-5c75d7c94b-5tcjp\" (UID: \"527a2a9e-741e-4f1b-8546-a852d73a836b\") " pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-5tcjp" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.392636 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrtfv\" (UniqueName: \"kubernetes.io/projected/4cc947bb-881d-4592-9db6-ea53f6694ea5-kube-api-access-mrtfv\") pod \"horizon-operator-controller-manager-5d86b44686-c776j\" (UID: \"4cc947bb-881d-4592-9db6-ea53f6694ea5\") " pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-c776j" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.392678 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bbd5\" (UniqueName: \"kubernetes.io/projected/6f79eaa5-12ee-4c47-8c5f-c0fa9190bc2d-kube-api-access-4bbd5\") pod \"designate-operator-controller-manager-56dfb6b67f-47bnx\" (UID: \"6f79eaa5-12ee-4c47-8c5f-c0fa9190bc2d\") " pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-47bnx" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.392698 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8swt\" (UniqueName: \"kubernetes.io/projected/50ff7c3e-8c4a-45e9-abe0-f5391a5c7c95-kube-api-access-c8swt\") pod \"manila-operator-controller-manager-7bb88cb858-qgvkm\" (UID: \"50ff7c3e-8c4a-45e9-abe0-f5391a5c7c95\") " pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-qgvkm" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.392726 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm85l\" (UniqueName: \"kubernetes.io/projected/82279b3c-bca1-4891-b7d9-a367005ad84e-kube-api-access-lm85l\") pod \"infra-operator-controller-manager-769d9c7585-qq8sm\" (UID: \"82279b3c-bca1-4891-b7d9-a367005ad84e\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-qq8sm" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.392747 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdvwc\" (UniqueName: \"kubernetes.io/projected/6935f82b-929a-421a-95d0-3315c9d0ef70-kube-api-access-bdvwc\") pod \"keystone-operator-controller-manager-7879fb76fd-rxpxw\" (UID: \"6935f82b-929a-421a-95d0-3315c9d0ef70\") " pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-rxpxw" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.392766 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbsg6\" (UniqueName: \"kubernetes.io/projected/042a59c3-a00e-4934-b16e-c731f0ab903f-kube-api-access-nbsg6\") pod \"glance-operator-controller-manager-8667fbf6f6-qfgkh\" (UID: \"042a59c3-a00e-4934-b16e-c731f0ab903f\") " pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-qfgkh" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.392785 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h42qr\" (UniqueName: \"kubernetes.io/projected/5b70e018-1ab3-4f23-b43f-1f433a096b81-kube-api-access-h42qr\") pod \"heat-operator-controller-manager-bf4c6585d-9tvg2\" (UID: \"5b70e018-1ab3-4f23-b43f-1f433a096b81\") " pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-9tvg2" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.406651 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-ks2b7" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.411152 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bbd5\" (UniqueName: \"kubernetes.io/projected/6f79eaa5-12ee-4c47-8c5f-c0fa9190bc2d-kube-api-access-4bbd5\") pod \"designate-operator-controller-manager-56dfb6b67f-47bnx\" (UID: \"6f79eaa5-12ee-4c47-8c5f-c0fa9190bc2d\") " pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-47bnx" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.411938 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h42qr\" (UniqueName: \"kubernetes.io/projected/5b70e018-1ab3-4f23-b43f-1f433a096b81-kube-api-access-h42qr\") pod \"heat-operator-controller-manager-bf4c6585d-9tvg2\" (UID: \"5b70e018-1ab3-4f23-b43f-1f433a096b81\") " pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-9tvg2" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.412065 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.420277 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-d78br" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.442832 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-6dc664666c-htkt7"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.443901 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbsg6\" (UniqueName: \"kubernetes.io/projected/042a59c3-a00e-4934-b16e-c731f0ab903f-kube-api-access-nbsg6\") pod \"glance-operator-controller-manager-8667fbf6f6-qfgkh\" (UID: \"042a59c3-a00e-4934-b16e-c731f0ab903f\") " pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-qfgkh" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.444374 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.447126 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.447242 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-htkt7" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.448021 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.448870 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-b5rkc" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.451553 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-7jfdq" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.459908 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-6dc664666c-htkt7"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.462600 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-47bnx" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.474385 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-799cb6ffd6-56qvk"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.475601 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-56qvk" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.478553 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-jw9n7" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.493902 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrtfv\" (UniqueName: \"kubernetes.io/projected/4cc947bb-881d-4592-9db6-ea53f6694ea5-kube-api-access-mrtfv\") pod \"horizon-operator-controller-manager-5d86b44686-c776j\" (UID: \"4cc947bb-881d-4592-9db6-ea53f6694ea5\") " pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-c776j" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.493950 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2srn\" (UniqueName: \"kubernetes.io/projected/9c496b90-00da-4367-90f5-347a0c5a8ac6-kube-api-access-q2srn\") pod \"ovn-operator-controller-manager-5bdf4f7f7f-8hrnp\" (UID: \"9c496b90-00da-4367-90f5-347a0c5a8ac6\") " pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-8hrnp" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.493984 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8swt\" (UniqueName: \"kubernetes.io/projected/50ff7c3e-8c4a-45e9-abe0-f5391a5c7c95-kube-api-access-c8swt\") pod \"manila-operator-controller-manager-7bb88cb858-qgvkm\" (UID: \"50ff7c3e-8c4a-45e9-abe0-f5391a5c7c95\") " pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-qgvkm" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.494005 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g84m5\" (UniqueName: \"kubernetes.io/projected/38f7bf7e-6d60-4cda-a988-ef9ce22883f8-kube-api-access-g84m5\") pod \"nova-operator-controller-manager-86d796d84d-6ht8s\" (UID: \"38f7bf7e-6d60-4cda-a988-ef9ce22883f8\") " pod="openstack-operators/nova-operator-controller-manager-86d796d84d-6ht8s" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.494039 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm85l\" (UniqueName: \"kubernetes.io/projected/82279b3c-bca1-4891-b7d9-a367005ad84e-kube-api-access-lm85l\") pod \"infra-operator-controller-manager-769d9c7585-qq8sm\" (UID: \"82279b3c-bca1-4891-b7d9-a367005ad84e\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-qq8sm" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.494061 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdvwc\" (UniqueName: \"kubernetes.io/projected/6935f82b-929a-421a-95d0-3315c9d0ef70-kube-api-access-bdvwc\") pod \"keystone-operator-controller-manager-7879fb76fd-rxpxw\" (UID: \"6935f82b-929a-421a-95d0-3315c9d0ef70\") " pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-rxpxw" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.494091 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgnhg\" (UniqueName: \"kubernetes.io/projected/a659f54b-26d9-4681-963a-40348401f023-kube-api-access-dgnhg\") pod \"octavia-operator-controller-manager-6fdc856c5d-bg99z\" (UID: \"a659f54b-26d9-4681-963a-40348401f023\") " pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-bg99z" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.494113 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7tfd\" (UniqueName: \"kubernetes.io/projected/7c53dd20-9115-4fac-a49e-b6b099150245-kube-api-access-w7tfd\") pod \"neutron-operator-controller-manager-66b7d6f598-hgbh8\" (UID: \"7c53dd20-9115-4fac-a49e-b6b099150245\") " pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-hgbh8" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.494135 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/82279b3c-bca1-4891-b7d9-a367005ad84e-cert\") pod \"infra-operator-controller-manager-769d9c7585-qq8sm\" (UID: \"82279b3c-bca1-4891-b7d9-a367005ad84e\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-qq8sm" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.494154 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-772zl\" (UniqueName: \"kubernetes.io/projected/527a2a9e-741e-4f1b-8546-a852d73a836b-kube-api-access-772zl\") pod \"ironic-operator-controller-manager-5c75d7c94b-5tcjp\" (UID: \"527a2a9e-741e-4f1b-8546-a852d73a836b\") " pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-5tcjp" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.494180 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkrj6\" (UniqueName: \"kubernetes.io/projected/620fe37e-07bb-4286-b1e8-fd62db1cf022-kube-api-access-hkrj6\") pod \"mariadb-operator-controller-manager-6f8c5b86cb-5rhv2\" (UID: \"620fe37e-07bb-4286-b1e8-fd62db1cf022\") " pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-5rhv2" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.498152 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/82279b3c-bca1-4891-b7d9-a367005ad84e-cert\") pod \"infra-operator-controller-manager-769d9c7585-qq8sm\" (UID: \"82279b3c-bca1-4891-b7d9-a367005ad84e\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-qq8sm" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.500930 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-qfgkh" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.510257 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-799cb6ffd6-56qvk"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.514121 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrtfv\" (UniqueName: \"kubernetes.io/projected/4cc947bb-881d-4592-9db6-ea53f6694ea5-kube-api-access-mrtfv\") pod \"horizon-operator-controller-manager-5d86b44686-c776j\" (UID: \"4cc947bb-881d-4592-9db6-ea53f6694ea5\") " pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-c776j" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.514509 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm85l\" (UniqueName: \"kubernetes.io/projected/82279b3c-bca1-4891-b7d9-a367005ad84e-kube-api-access-lm85l\") pod \"infra-operator-controller-manager-769d9c7585-qq8sm\" (UID: \"82279b3c-bca1-4891-b7d9-a367005ad84e\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-qq8sm" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.517501 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8swt\" (UniqueName: \"kubernetes.io/projected/50ff7c3e-8c4a-45e9-abe0-f5391a5c7c95-kube-api-access-c8swt\") pod \"manila-operator-controller-manager-7bb88cb858-qgvkm\" (UID: \"50ff7c3e-8c4a-45e9-abe0-f5391a5c7c95\") " pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-qgvkm" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.517580 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-9tvg2" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.520244 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-772zl\" (UniqueName: \"kubernetes.io/projected/527a2a9e-741e-4f1b-8546-a852d73a836b-kube-api-access-772zl\") pod \"ironic-operator-controller-manager-5c75d7c94b-5tcjp\" (UID: \"527a2a9e-741e-4f1b-8546-a852d73a836b\") " pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-5tcjp" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.522364 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdvwc\" (UniqueName: \"kubernetes.io/projected/6935f82b-929a-421a-95d0-3315c9d0ef70-kube-api-access-bdvwc\") pod \"keystone-operator-controller-manager-7879fb76fd-rxpxw\" (UID: \"6935f82b-929a-421a-95d0-3315c9d0ef70\") " pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-rxpxw" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.528706 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7798859c74-62tqn"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.531939 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-62tqn" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.533759 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-xhrx7" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.540118 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-c776j" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.540776 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7798859c74-62tqn"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.551269 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-qq8sm" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.564943 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-5tcjp" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.581904 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-8464cf66df-c2gjt"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.583588 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-8464cf66df-c2gjt" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.586372 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-wbvtp" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.594033 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-rxpxw" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.596812 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-8464cf66df-c2gjt"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.600216 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkrj6\" (UniqueName: \"kubernetes.io/projected/620fe37e-07bb-4286-b1e8-fd62db1cf022-kube-api-access-hkrj6\") pod \"mariadb-operator-controller-manager-6f8c5b86cb-5rhv2\" (UID: \"620fe37e-07bb-4286-b1e8-fd62db1cf022\") " pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-5rhv2" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.600289 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2srn\" (UniqueName: \"kubernetes.io/projected/9c496b90-00da-4367-90f5-347a0c5a8ac6-kube-api-access-q2srn\") pod \"ovn-operator-controller-manager-5bdf4f7f7f-8hrnp\" (UID: \"9c496b90-00da-4367-90f5-347a0c5a8ac6\") " pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-8hrnp" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.600321 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwd2w\" (UniqueName: \"kubernetes.io/projected/b5998049-f03f-4de4-9601-dd970e3215fe-kube-api-access-fwd2w\") pod \"swift-operator-controller-manager-799cb6ffd6-56qvk\" (UID: \"b5998049-f03f-4de4-9601-dd970e3215fe\") " pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-56qvk" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.600378 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g84m5\" (UniqueName: \"kubernetes.io/projected/38f7bf7e-6d60-4cda-a988-ef9ce22883f8-kube-api-access-g84m5\") pod \"nova-operator-controller-manager-86d796d84d-6ht8s\" (UID: \"38f7bf7e-6d60-4cda-a988-ef9ce22883f8\") " pod="openstack-operators/nova-operator-controller-manager-86d796d84d-6ht8s" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.600416 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/14445d59-81da-4b26-bd5d-2bbe2bd1a9c0-cert\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r\" (UID: \"14445d59-81da-4b26-bd5d-2bbe2bd1a9c0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.600437 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j89t6\" (UniqueName: \"kubernetes.io/projected/c5127c1b-7342-4669-b04b-8484fd2ac326-kube-api-access-j89t6\") pod \"placement-operator-controller-manager-6dc664666c-htkt7\" (UID: \"c5127c1b-7342-4669-b04b-8484fd2ac326\") " pod="openstack-operators/placement-operator-controller-manager-6dc664666c-htkt7" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.600523 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgnhg\" (UniqueName: \"kubernetes.io/projected/a659f54b-26d9-4681-963a-40348401f023-kube-api-access-dgnhg\") pod \"octavia-operator-controller-manager-6fdc856c5d-bg99z\" (UID: \"a659f54b-26d9-4681-963a-40348401f023\") " pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-bg99z" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.600550 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7tfd\" (UniqueName: \"kubernetes.io/projected/7c53dd20-9115-4fac-a49e-b6b099150245-kube-api-access-w7tfd\") pod \"neutron-operator-controller-manager-66b7d6f598-hgbh8\" (UID: \"7c53dd20-9115-4fac-a49e-b6b099150245\") " pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-hgbh8" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.600583 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tjcx\" (UniqueName: \"kubernetes.io/projected/14445d59-81da-4b26-bd5d-2bbe2bd1a9c0-kube-api-access-9tjcx\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r\" (UID: \"14445d59-81da-4b26-bd5d-2bbe2bd1a9c0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.603074 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-qgvkm" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.620613 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2srn\" (UniqueName: \"kubernetes.io/projected/9c496b90-00da-4367-90f5-347a0c5a8ac6-kube-api-access-q2srn\") pod \"ovn-operator-controller-manager-5bdf4f7f7f-8hrnp\" (UID: \"9c496b90-00da-4367-90f5-347a0c5a8ac6\") " pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-8hrnp" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.620818 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7tfd\" (UniqueName: \"kubernetes.io/projected/7c53dd20-9115-4fac-a49e-b6b099150245-kube-api-access-w7tfd\") pod \"neutron-operator-controller-manager-66b7d6f598-hgbh8\" (UID: \"7c53dd20-9115-4fac-a49e-b6b099150245\") " pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-hgbh8" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.621192 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g84m5\" (UniqueName: \"kubernetes.io/projected/38f7bf7e-6d60-4cda-a988-ef9ce22883f8-kube-api-access-g84m5\") pod \"nova-operator-controller-manager-86d796d84d-6ht8s\" (UID: \"38f7bf7e-6d60-4cda-a988-ef9ce22883f8\") " pod="openstack-operators/nova-operator-controller-manager-86d796d84d-6ht8s" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.621202 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgnhg\" (UniqueName: \"kubernetes.io/projected/a659f54b-26d9-4681-963a-40348401f023-kube-api-access-dgnhg\") pod \"octavia-operator-controller-manager-6fdc856c5d-bg99z\" (UID: \"a659f54b-26d9-4681-963a-40348401f023\") " pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-bg99z" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.622026 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkrj6\" (UniqueName: \"kubernetes.io/projected/620fe37e-07bb-4286-b1e8-fd62db1cf022-kube-api-access-hkrj6\") pod \"mariadb-operator-controller-manager-6f8c5b86cb-5rhv2\" (UID: \"620fe37e-07bb-4286-b1e8-fd62db1cf022\") " pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-5rhv2" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.661218 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-5rhv2" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.667696 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-hgbh8" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.680372 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-6ht8s" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.682205 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-bg99z" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.704000 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-8hrnp" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.704932 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwd2w\" (UniqueName: \"kubernetes.io/projected/b5998049-f03f-4de4-9601-dd970e3215fe-kube-api-access-fwd2w\") pod \"swift-operator-controller-manager-799cb6ffd6-56qvk\" (UID: \"b5998049-f03f-4de4-9601-dd970e3215fe\") " pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-56qvk" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.704982 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn7m5\" (UniqueName: \"kubernetes.io/projected/2463fe89-bc43-469f-9837-3ffd4b75605c-kube-api-access-fn7m5\") pod \"test-operator-controller-manager-8464cf66df-c2gjt\" (UID: \"2463fe89-bc43-469f-9837-3ffd4b75605c\") " pod="openstack-operators/test-operator-controller-manager-8464cf66df-c2gjt" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.705006 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/14445d59-81da-4b26-bd5d-2bbe2bd1a9c0-cert\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r\" (UID: \"14445d59-81da-4b26-bd5d-2bbe2bd1a9c0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.705026 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j89t6\" (UniqueName: \"kubernetes.io/projected/c5127c1b-7342-4669-b04b-8484fd2ac326-kube-api-access-j89t6\") pod \"placement-operator-controller-manager-6dc664666c-htkt7\" (UID: \"c5127c1b-7342-4669-b04b-8484fd2ac326\") " pod="openstack-operators/placement-operator-controller-manager-6dc664666c-htkt7" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.705055 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk2q6\" (UniqueName: \"kubernetes.io/projected/6486d960-766f-4622-bad1-853e2b6a3fce-kube-api-access-qk2q6\") pod \"telemetry-operator-controller-manager-7798859c74-62tqn\" (UID: \"6486d960-766f-4622-bad1-853e2b6a3fce\") " pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-62tqn" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.705103 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tjcx\" (UniqueName: \"kubernetes.io/projected/14445d59-81da-4b26-bd5d-2bbe2bd1a9c0-kube-api-access-9tjcx\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r\" (UID: \"14445d59-81da-4b26-bd5d-2bbe2bd1a9c0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r" Nov 23 06:56:55 crc kubenswrapper[4559]: E1123 06:56:55.705544 4559 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 23 06:56:55 crc kubenswrapper[4559]: E1123 06:56:55.705580 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/14445d59-81da-4b26-bd5d-2bbe2bd1a9c0-cert podName:14445d59-81da-4b26-bd5d-2bbe2bd1a9c0 nodeName:}" failed. No retries permitted until 2025-11-23 06:56:56.205568564 +0000 UTC m=+718.227554178 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/14445d59-81da-4b26-bd5d-2bbe2bd1a9c0-cert") pod "openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r" (UID: "14445d59-81da-4b26-bd5d-2bbe2bd1a9c0") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.707369 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7brkl"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.708341 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7brkl" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.710281 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-6b56x" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.726537 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7brkl"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.734668 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tjcx\" (UniqueName: \"kubernetes.io/projected/14445d59-81da-4b26-bd5d-2bbe2bd1a9c0-kube-api-access-9tjcx\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r\" (UID: \"14445d59-81da-4b26-bd5d-2bbe2bd1a9c0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.736021 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j89t6\" (UniqueName: \"kubernetes.io/projected/c5127c1b-7342-4669-b04b-8484fd2ac326-kube-api-access-j89t6\") pod \"placement-operator-controller-manager-6dc664666c-htkt7\" (UID: \"c5127c1b-7342-4669-b04b-8484fd2ac326\") " pod="openstack-operators/placement-operator-controller-manager-6dc664666c-htkt7" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.754988 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwd2w\" (UniqueName: \"kubernetes.io/projected/b5998049-f03f-4de4-9601-dd970e3215fe-kube-api-access-fwd2w\") pod \"swift-operator-controller-manager-799cb6ffd6-56qvk\" (UID: \"b5998049-f03f-4de4-9601-dd970e3215fe\") " pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-56qvk" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.780236 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-htkt7" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.805833 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-56qvk" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.812812 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cc7td\" (UniqueName: \"kubernetes.io/projected/a0f57f27-60d8-4fb8-ba76-c4ea5f3b99f6-kube-api-access-cc7td\") pod \"watcher-operator-controller-manager-7cd4fb6f79-7brkl\" (UID: \"a0f57f27-60d8-4fb8-ba76-c4ea5f3b99f6\") " pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7brkl" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.812909 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn7m5\" (UniqueName: \"kubernetes.io/projected/2463fe89-bc43-469f-9837-3ffd4b75605c-kube-api-access-fn7m5\") pod \"test-operator-controller-manager-8464cf66df-c2gjt\" (UID: \"2463fe89-bc43-469f-9837-3ffd4b75605c\") " pod="openstack-operators/test-operator-controller-manager-8464cf66df-c2gjt" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.812961 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk2q6\" (UniqueName: \"kubernetes.io/projected/6486d960-766f-4622-bad1-853e2b6a3fce-kube-api-access-qk2q6\") pod \"telemetry-operator-controller-manager-7798859c74-62tqn\" (UID: \"6486d960-766f-4622-bad1-853e2b6a3fce\") " pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-62tqn" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.838176 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn7m5\" (UniqueName: \"kubernetes.io/projected/2463fe89-bc43-469f-9837-3ffd4b75605c-kube-api-access-fn7m5\") pod \"test-operator-controller-manager-8464cf66df-c2gjt\" (UID: \"2463fe89-bc43-469f-9837-3ffd4b75605c\") " pod="openstack-operators/test-operator-controller-manager-8464cf66df-c2gjt" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.843611 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk2q6\" (UniqueName: \"kubernetes.io/projected/6486d960-766f-4622-bad1-853e2b6a3fce-kube-api-access-qk2q6\") pod \"telemetry-operator-controller-manager-7798859c74-62tqn\" (UID: \"6486d960-766f-4622-bad1-853e2b6a3fce\") " pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-62tqn" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.848621 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-62tqn" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.880053 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-gmv99"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.880976 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-gmv99" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.892365 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.892826 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-zwbj8" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.894981 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-gmv99"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.905073 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-8464cf66df-c2gjt" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.916875 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cc7td\" (UniqueName: \"kubernetes.io/projected/a0f57f27-60d8-4fb8-ba76-c4ea5f3b99f6-kube-api-access-cc7td\") pod \"watcher-operator-controller-manager-7cd4fb6f79-7brkl\" (UID: \"a0f57f27-60d8-4fb8-ba76-c4ea5f3b99f6\") " pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7brkl" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.920234 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7768f8c84f-ks2b7"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.949745 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-8667fbf6f6-qfgkh"] Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.960042 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cc7td\" (UniqueName: \"kubernetes.io/projected/a0f57f27-60d8-4fb8-ba76-c4ea5f3b99f6-kube-api-access-cc7td\") pod \"watcher-operator-controller-manager-7cd4fb6f79-7brkl\" (UID: \"a0f57f27-60d8-4fb8-ba76-c4ea5f3b99f6\") " pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7brkl" Nov 23 06:56:55 crc kubenswrapper[4559]: I1123 06:56:55.979536 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-56dfb6b67f-47bnx"] Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.000470 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f6s9r"] Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.001548 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f6s9r" Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.004071 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f6s9r"] Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.006013 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-jfgxk" Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.018549 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-d78br"] Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.022329 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ec988465-9f4e-46fe-beb3-e9032bf589d0-cert\") pod \"openstack-operator-controller-manager-6cb9dc54f8-gmv99\" (UID: \"ec988465-9f4e-46fe-beb3-e9032bf589d0\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-gmv99" Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.022372 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nlqt\" (UniqueName: \"kubernetes.io/projected/ec988465-9f4e-46fe-beb3-e9032bf589d0-kube-api-access-7nlqt\") pod \"openstack-operator-controller-manager-6cb9dc54f8-gmv99\" (UID: \"ec988465-9f4e-46fe-beb3-e9032bf589d0\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-gmv99" Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.063174 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7brkl" Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.124888 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ec988465-9f4e-46fe-beb3-e9032bf589d0-cert\") pod \"openstack-operator-controller-manager-6cb9dc54f8-gmv99\" (UID: \"ec988465-9f4e-46fe-beb3-e9032bf589d0\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-gmv99" Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.124930 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nlqt\" (UniqueName: \"kubernetes.io/projected/ec988465-9f4e-46fe-beb3-e9032bf589d0-kube-api-access-7nlqt\") pod \"openstack-operator-controller-manager-6cb9dc54f8-gmv99\" (UID: \"ec988465-9f4e-46fe-beb3-e9032bf589d0\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-gmv99" Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.125006 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4psm4\" (UniqueName: \"kubernetes.io/projected/465518b5-58ac-4347-aae0-94b8f68a6b36-kube-api-access-4psm4\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-f6s9r\" (UID: \"465518b5-58ac-4347-aae0-94b8f68a6b36\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f6s9r" Nov 23 06:56:56 crc kubenswrapper[4559]: E1123 06:56:56.125203 4559 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 23 06:56:56 crc kubenswrapper[4559]: E1123 06:56:56.125245 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ec988465-9f4e-46fe-beb3-e9032bf589d0-cert podName:ec988465-9f4e-46fe-beb3-e9032bf589d0 nodeName:}" failed. No retries permitted until 2025-11-23 06:56:56.62523148 +0000 UTC m=+718.647217084 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ec988465-9f4e-46fe-beb3-e9032bf589d0-cert") pod "openstack-operator-controller-manager-6cb9dc54f8-gmv99" (UID: "ec988465-9f4e-46fe-beb3-e9032bf589d0") : secret "webhook-server-cert" not found Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.167071 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.173924 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.192621 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nlqt\" (UniqueName: \"kubernetes.io/projected/ec988465-9f4e-46fe-beb3-e9032bf589d0-kube-api-access-7nlqt\") pod \"openstack-operator-controller-manager-6cb9dc54f8-gmv99\" (UID: \"ec988465-9f4e-46fe-beb3-e9032bf589d0\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-gmv99" Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.227540 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/14445d59-81da-4b26-bd5d-2bbe2bd1a9c0-cert\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r\" (UID: \"14445d59-81da-4b26-bd5d-2bbe2bd1a9c0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r" Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.227612 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4psm4\" (UniqueName: \"kubernetes.io/projected/465518b5-58ac-4347-aae0-94b8f68a6b36-kube-api-access-4psm4\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-f6s9r\" (UID: \"465518b5-58ac-4347-aae0-94b8f68a6b36\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f6s9r" Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.234705 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/14445d59-81da-4b26-bd5d-2bbe2bd1a9c0-cert\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r\" (UID: \"14445d59-81da-4b26-bd5d-2bbe2bd1a9c0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r" Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.241454 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4psm4\" (UniqueName: \"kubernetes.io/projected/465518b5-58ac-4347-aae0-94b8f68a6b36-kube-api-access-4psm4\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-f6s9r\" (UID: \"465518b5-58ac-4347-aae0-94b8f68a6b36\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f6s9r" Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.340863 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f6s9r" Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.366698 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r" Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.436481 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-769d9c7585-qq8sm"] Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.441845 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-bf4c6585d-9tvg2"] Nov 23 06:56:56 crc kubenswrapper[4559]: W1123 06:56:56.447799 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b70e018_1ab3_4f23_b43f_1f433a096b81.slice/crio-3be1c215207a25d0908f60deaff249b01c9dabe1aec38c2cfed9eb683bc6e7ed WatchSource:0}: Error finding container 3be1c215207a25d0908f60deaff249b01c9dabe1aec38c2cfed9eb683bc6e7ed: Status 404 returned error can't find the container with id 3be1c215207a25d0908f60deaff249b01c9dabe1aec38c2cfed9eb683bc6e7ed Nov 23 06:56:56 crc kubenswrapper[4559]: W1123 06:56:56.450798 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82279b3c_bca1_4891_b7d9_a367005ad84e.slice/crio-1bd920b5406f3c51f6f56d4e8619d4b7a54e8e8a0a24e96d93b84c220f07b222 WatchSource:0}: Error finding container 1bd920b5406f3c51f6f56d4e8619d4b7a54e8e8a0a24e96d93b84c220f07b222: Status 404 returned error can't find the container with id 1bd920b5406f3c51f6f56d4e8619d4b7a54e8e8a0a24e96d93b84c220f07b222 Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.618876 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5c75d7c94b-5tcjp"] Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.624350 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7879fb76fd-rxpxw"] Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.629543 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6fdc856c5d-bg99z"] Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.635713 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7bb88cb858-qgvkm"] Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.635965 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ec988465-9f4e-46fe-beb3-e9032bf589d0-cert\") pod \"openstack-operator-controller-manager-6cb9dc54f8-gmv99\" (UID: \"ec988465-9f4e-46fe-beb3-e9032bf589d0\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-gmv99" Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.640923 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-5rhv2"] Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.642374 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ec988465-9f4e-46fe-beb3-e9032bf589d0-cert\") pod \"openstack-operator-controller-manager-6cb9dc54f8-gmv99\" (UID: \"ec988465-9f4e-46fe-beb3-e9032bf589d0\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-gmv99" Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.650175 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d86b44686-c776j"] Nov 23 06:56:56 crc kubenswrapper[4559]: W1123 06:56:56.658480 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50ff7c3e_8c4a_45e9_abe0_f5391a5c7c95.slice/crio-ac542d8d1ec76f2486fc67557f1a4792e4fe8148bb66224c266cec03074ceca3 WatchSource:0}: Error finding container ac542d8d1ec76f2486fc67557f1a4792e4fe8148bb66224c266cec03074ceca3: Status 404 returned error can't find the container with id ac542d8d1ec76f2486fc67557f1a4792e4fe8148bb66224c266cec03074ceca3 Nov 23 06:56:56 crc kubenswrapper[4559]: W1123 06:56:56.658810 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4cc947bb_881d_4592_9db6_ea53f6694ea5.slice/crio-bd137a6ee3aa038319a24421f8c6563e775e02a10457b658a2c74125bba2381c WatchSource:0}: Error finding container bd137a6ee3aa038319a24421f8c6563e775e02a10457b658a2c74125bba2381c: Status 404 returned error can't find the container with id bd137a6ee3aa038319a24421f8c6563e775e02a10457b658a2c74125bba2381c Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.786887 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-8hrnp"] Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.791933 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-8464cf66df-c2gjt"] Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.792940 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-66b7d6f598-hgbh8"] Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.815116 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7798859c74-62tqn"] Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.816142 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7brkl"] Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.818598 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-gmv99" Nov 23 06:56:56 crc kubenswrapper[4559]: E1123 06:56:56.820508 4559 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fn7m5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-8464cf66df-c2gjt_openstack-operators(2463fe89-bc43-469f-9837-3ffd4b75605c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.824637 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-6dc664666c-htkt7"] Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.825804 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-86d796d84d-6ht8s"] Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.829137 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-799cb6ffd6-56qvk"] Nov 23 06:56:56 crc kubenswrapper[4559]: E1123 06:56:56.835192 4559 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j89t6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-6dc664666c-htkt7_openstack-operators(c5127c1b-7342-4669-b04b-8484fd2ac326): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 23 06:56:56 crc kubenswrapper[4559]: E1123 06:56:56.835393 4559 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q2srn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-5bdf4f7f7f-8hrnp_openstack-operators(9c496b90-00da-4367-90f5-347a0c5a8ac6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 23 06:56:56 crc kubenswrapper[4559]: E1123 06:56:56.839511 4559 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qk2q6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-7798859c74-62tqn_openstack-operators(6486d960-766f-4622-bad1-853e2b6a3fce): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 23 06:56:56 crc kubenswrapper[4559]: W1123 06:56:56.840780 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda0f57f27_60d8_4fb8_ba76_c4ea5f3b99f6.slice/crio-17ad7729e99851b897fd1a55c385dd935aa4a47a1c5f1eb4d3f51eb6a6055a77 WatchSource:0}: Error finding container 17ad7729e99851b897fd1a55c385dd935aa4a47a1c5f1eb4d3f51eb6a6055a77: Status 404 returned error can't find the container with id 17ad7729e99851b897fd1a55c385dd935aa4a47a1c5f1eb4d3f51eb6a6055a77 Nov 23 06:56:56 crc kubenswrapper[4559]: W1123 06:56:56.845022 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38f7bf7e_6d60_4cda_a988_ef9ce22883f8.slice/crio-2ab8b5b22a5658f23fdd232e267e671b7ef24aac83fb51ecfddd438781c1d67c WatchSource:0}: Error finding container 2ab8b5b22a5658f23fdd232e267e671b7ef24aac83fb51ecfddd438781c1d67c: Status 404 returned error can't find the container with id 2ab8b5b22a5658f23fdd232e267e671b7ef24aac83fb51ecfddd438781c1d67c Nov 23 06:56:56 crc kubenswrapper[4559]: E1123 06:56:56.849710 4559 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cc7td,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-7cd4fb6f79-7brkl_openstack-operators(a0f57f27-60d8-4fb8-ba76-c4ea5f3b99f6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 23 06:56:56 crc kubenswrapper[4559]: W1123 06:56:56.850264 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5998049_f03f_4de4_9601_dd970e3215fe.slice/crio-62a849fcfefba5c4d29cd51532d8f37479d36f3c2985bba29e3ebd003ced4581 WatchSource:0}: Error finding container 62a849fcfefba5c4d29cd51532d8f37479d36f3c2985bba29e3ebd003ced4581: Status 404 returned error can't find the container with id 62a849fcfefba5c4d29cd51532d8f37479d36f3c2985bba29e3ebd003ced4581 Nov 23 06:56:56 crc kubenswrapper[4559]: E1123 06:56:56.855347 4559 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-g84m5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-86d796d84d-6ht8s_openstack-operators(38f7bf7e-6d60-4cda-a988-ef9ce22883f8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 23 06:56:56 crc kubenswrapper[4559]: E1123 06:56:56.855525 4559 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fwd2w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-799cb6ffd6-56qvk_openstack-operators(b5998049-f03f-4de4-9601-dd970e3215fe): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.967128 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f6s9r"] Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.970851 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-8hrnp" event={"ID":"9c496b90-00da-4367-90f5-347a0c5a8ac6","Type":"ContainerStarted","Data":"d33eda18d400015dc9e21acc91ee6fb6f0c7f403f54ab99d8244ddc9c3ae51be"} Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.974864 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7brkl" event={"ID":"a0f57f27-60d8-4fb8-ba76-c4ea5f3b99f6","Type":"ContainerStarted","Data":"17ad7729e99851b897fd1a55c385dd935aa4a47a1c5f1eb4d3f51eb6a6055a77"} Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.979415 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r"] Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.980200 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-d78br" event={"ID":"35944557-7948-4d54-8225-788d51eb01d1","Type":"ContainerStarted","Data":"238e739d569a947d7c94d74b91f2ce252a881bbf8c9d3cc35fcfa3de99d72542"} Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.983694 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-hgbh8" event={"ID":"7c53dd20-9115-4fac-a49e-b6b099150245","Type":"ContainerStarted","Data":"b1fb437d9c92701dc20f2d64fec854c95b3c2e780db332d2852d60d451c7b84f"} Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.984994 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-8464cf66df-c2gjt" event={"ID":"2463fe89-bc43-469f-9837-3ffd4b75605c","Type":"ContainerStarted","Data":"2b23784b31f79acfe81219b2f1e8dcf7044cd6d9ef362932a13f4de3ee614e3a"} Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.988740 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-ks2b7" event={"ID":"090f1d86-b4e0-43ca-9aee-49c771cab8c1","Type":"ContainerStarted","Data":"8b3fef6f6d1ae41c37df6803e5c9c64a75a2df53a9a931f00ea2322af91b5a05"} Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.991123 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-56qvk" event={"ID":"b5998049-f03f-4de4-9601-dd970e3215fe","Type":"ContainerStarted","Data":"62a849fcfefba5c4d29cd51532d8f37479d36f3c2985bba29e3ebd003ced4581"} Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.992238 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-qgvkm" event={"ID":"50ff7c3e-8c4a-45e9-abe0-f5391a5c7c95","Type":"ContainerStarted","Data":"ac542d8d1ec76f2486fc67557f1a4792e4fe8148bb66224c266cec03074ceca3"} Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.993220 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-9tvg2" event={"ID":"5b70e018-1ab3-4f23-b43f-1f433a096b81","Type":"ContainerStarted","Data":"3be1c215207a25d0908f60deaff249b01c9dabe1aec38c2cfed9eb683bc6e7ed"} Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.994445 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-5tcjp" event={"ID":"527a2a9e-741e-4f1b-8546-a852d73a836b","Type":"ContainerStarted","Data":"c3a6ff6860ec64a5edf8cbd1ecb9f54d5601d16234024026e9840e552f96b352"} Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.995804 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-qq8sm" event={"ID":"82279b3c-bca1-4891-b7d9-a367005ad84e","Type":"ContainerStarted","Data":"1bd920b5406f3c51f6f56d4e8619d4b7a54e8e8a0a24e96d93b84c220f07b222"} Nov 23 06:56:56 crc kubenswrapper[4559]: I1123 06:56:56.998231 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-6ht8s" event={"ID":"38f7bf7e-6d60-4cda-a988-ef9ce22883f8","Type":"ContainerStarted","Data":"2ab8b5b22a5658f23fdd232e267e671b7ef24aac83fb51ecfddd438781c1d67c"} Nov 23 06:56:57 crc kubenswrapper[4559]: I1123 06:56:57.000085 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-c776j" event={"ID":"4cc947bb-881d-4592-9db6-ea53f6694ea5","Type":"ContainerStarted","Data":"bd137a6ee3aa038319a24421f8c6563e775e02a10457b658a2c74125bba2381c"} Nov 23 06:56:57 crc kubenswrapper[4559]: I1123 06:56:57.001047 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-qfgkh" event={"ID":"042a59c3-a00e-4934-b16e-c731f0ab903f","Type":"ContainerStarted","Data":"4984e95b9fb017512195c3c3a487b80c4771f45bbaeb7ddb67ec182f3f238c2a"} Nov 23 06:56:57 crc kubenswrapper[4559]: W1123 06:56:57.014811 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod465518b5_58ac_4347_aae0_94b8f68a6b36.slice/crio-47892b436b2db37ec0ba8fcabba546706a985943bfb9bfa430ab2c0630963817 WatchSource:0}: Error finding container 47892b436b2db37ec0ba8fcabba546706a985943bfb9bfa430ab2c0630963817: Status 404 returned error can't find the container with id 47892b436b2db37ec0ba8fcabba546706a985943bfb9bfa430ab2c0630963817 Nov 23 06:56:57 crc kubenswrapper[4559]: I1123 06:56:57.015227 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-rxpxw" event={"ID":"6935f82b-929a-421a-95d0-3315c9d0ef70","Type":"ContainerStarted","Data":"839f66ec1b38783138d9bfa0c3ca3267cd2c9674ad179d1552c081e87481c1eb"} Nov 23 06:56:57 crc kubenswrapper[4559]: E1123 06:56:57.016016 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-8464cf66df-c2gjt" podUID="2463fe89-bc43-469f-9837-3ffd4b75605c" Nov 23 06:56:57 crc kubenswrapper[4559]: E1123 06:56:57.016355 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-htkt7" podUID="c5127c1b-7342-4669-b04b-8484fd2ac326" Nov 23 06:56:57 crc kubenswrapper[4559]: I1123 06:56:57.016910 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-bg99z" event={"ID":"a659f54b-26d9-4681-963a-40348401f023","Type":"ContainerStarted","Data":"b572357b7deaac0ece534e4f88d680cf917e83b33e6ecf04e8854bc149c117bd"} Nov 23 06:56:57 crc kubenswrapper[4559]: W1123 06:56:57.019397 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod14445d59_81da_4b26_bd5d_2bbe2bd1a9c0.slice/crio-00276327a09a7ac8d927f75d42a0a0cdd1403967766cc2642a18f20a1ad49afc WatchSource:0}: Error finding container 00276327a09a7ac8d927f75d42a0a0cdd1403967766cc2642a18f20a1ad49afc: Status 404 returned error can't find the container with id 00276327a09a7ac8d927f75d42a0a0cdd1403967766cc2642a18f20a1ad49afc Nov 23 06:56:57 crc kubenswrapper[4559]: I1123 06:56:57.020173 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-5rhv2" event={"ID":"620fe37e-07bb-4286-b1e8-fd62db1cf022","Type":"ContainerStarted","Data":"81d27a9123bde4ec9127ee9d9178c6def1c8f4bec435ac5395760c2fa54186ff"} Nov 23 06:56:57 crc kubenswrapper[4559]: I1123 06:56:57.022493 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-htkt7" event={"ID":"c5127c1b-7342-4669-b04b-8484fd2ac326","Type":"ContainerStarted","Data":"f3481c8bd4fdd72abe50292ec9202176284cc4e9954bbae2c52a0839bcb00565"} Nov 23 06:56:57 crc kubenswrapper[4559]: E1123 06:56:57.024560 4559 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4psm4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-f6s9r_openstack-operators(465518b5-58ac-4347-aae0-94b8f68a6b36): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 23 06:56:57 crc kubenswrapper[4559]: E1123 06:56:57.025035 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c\\\"\"" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-htkt7" podUID="c5127c1b-7342-4669-b04b-8484fd2ac326" Nov 23 06:56:57 crc kubenswrapper[4559]: I1123 06:56:57.025442 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-47bnx" event={"ID":"6f79eaa5-12ee-4c47-8c5f-c0fa9190bc2d","Type":"ContainerStarted","Data":"74d224072441e95f5fd60872f81288a69fefc1e3a8b1158bcc38dbb3ba48eeb4"} Nov 23 06:56:57 crc kubenswrapper[4559]: E1123 06:56:57.025809 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f6s9r" podUID="465518b5-58ac-4347-aae0-94b8f68a6b36" Nov 23 06:56:57 crc kubenswrapper[4559]: I1123 06:56:57.026955 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-62tqn" event={"ID":"6486d960-766f-4622-bad1-853e2b6a3fce","Type":"ContainerStarted","Data":"647f4e2994f6372f443080f886c83caf1e124167eb2daca479dfa645b0283531"} Nov 23 06:56:57 crc kubenswrapper[4559]: E1123 06:56:57.048404 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-8hrnp" podUID="9c496b90-00da-4367-90f5-347a0c5a8ac6" Nov 23 06:56:57 crc kubenswrapper[4559]: E1123 06:56:57.048817 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7brkl" podUID="a0f57f27-60d8-4fb8-ba76-c4ea5f3b99f6" Nov 23 06:56:57 crc kubenswrapper[4559]: E1123 06:56:57.137956 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-56qvk" podUID="b5998049-f03f-4de4-9601-dd970e3215fe" Nov 23 06:56:57 crc kubenswrapper[4559]: E1123 06:56:57.138080 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-62tqn" podUID="6486d960-766f-4622-bad1-853e2b6a3fce" Nov 23 06:56:57 crc kubenswrapper[4559]: E1123 06:56:57.165834 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-6ht8s" podUID="38f7bf7e-6d60-4cda-a988-ef9ce22883f8" Nov 23 06:56:57 crc kubenswrapper[4559]: I1123 06:56:57.256231 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-gmv99"] Nov 23 06:56:57 crc kubenswrapper[4559]: W1123 06:56:57.281051 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podec988465_9f4e_46fe_beb3_e9032bf589d0.slice/crio-19233b5757996599e3eee39a3aca8baa6b030f686414ecd52b4a7eebe2096d60 WatchSource:0}: Error finding container 19233b5757996599e3eee39a3aca8baa6b030f686414ecd52b4a7eebe2096d60: Status 404 returned error can't find the container with id 19233b5757996599e3eee39a3aca8baa6b030f686414ecd52b4a7eebe2096d60 Nov 23 06:56:58 crc kubenswrapper[4559]: I1123 06:56:58.050717 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-6ht8s" event={"ID":"38f7bf7e-6d60-4cda-a988-ef9ce22883f8","Type":"ContainerStarted","Data":"09d17bc7bedd24ce39abe2eb117c965f0b1723de1b71690567d455d2ed1b3869"} Nov 23 06:56:58 crc kubenswrapper[4559]: E1123 06:56:58.052946 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\"" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-6ht8s" podUID="38f7bf7e-6d60-4cda-a988-ef9ce22883f8" Nov 23 06:56:58 crc kubenswrapper[4559]: I1123 06:56:58.056951 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-8464cf66df-c2gjt" event={"ID":"2463fe89-bc43-469f-9837-3ffd4b75605c","Type":"ContainerStarted","Data":"8d8a52c4438471904f5b824a5b4fd9fb9bc9b96bd019a8fd5b28d715e4de3bd6"} Nov 23 06:56:58 crc kubenswrapper[4559]: I1123 06:56:58.058796 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r" event={"ID":"14445d59-81da-4b26-bd5d-2bbe2bd1a9c0","Type":"ContainerStarted","Data":"00276327a09a7ac8d927f75d42a0a0cdd1403967766cc2642a18f20a1ad49afc"} Nov 23 06:56:58 crc kubenswrapper[4559]: E1123 06:56:58.062105 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-8464cf66df-c2gjt" podUID="2463fe89-bc43-469f-9837-3ffd4b75605c" Nov 23 06:56:58 crc kubenswrapper[4559]: I1123 06:56:58.080216 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-htkt7" event={"ID":"c5127c1b-7342-4669-b04b-8484fd2ac326","Type":"ContainerStarted","Data":"029028e85ddd5eefc3cbe26fdef42a9dcbc00ad4e466effc2d7b669260722e71"} Nov 23 06:56:58 crc kubenswrapper[4559]: E1123 06:56:58.082843 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c\\\"\"" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-htkt7" podUID="c5127c1b-7342-4669-b04b-8484fd2ac326" Nov 23 06:56:58 crc kubenswrapper[4559]: I1123 06:56:58.097267 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-62tqn" event={"ID":"6486d960-766f-4622-bad1-853e2b6a3fce","Type":"ContainerStarted","Data":"a671981b5af98550fd0671e36171f2c78ae03a704adcb5937b7c74c5dfa78fea"} Nov 23 06:56:58 crc kubenswrapper[4559]: E1123 06:56:58.111418 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-62tqn" podUID="6486d960-766f-4622-bad1-853e2b6a3fce" Nov 23 06:56:58 crc kubenswrapper[4559]: I1123 06:56:58.115138 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-8hrnp" event={"ID":"9c496b90-00da-4367-90f5-347a0c5a8ac6","Type":"ContainerStarted","Data":"bea0852bf89d741692373fb95994bdae42166a1480736c4f5a48cd509368dc3d"} Nov 23 06:56:58 crc kubenswrapper[4559]: E1123 06:56:58.118033 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-8hrnp" podUID="9c496b90-00da-4367-90f5-347a0c5a8ac6" Nov 23 06:56:58 crc kubenswrapper[4559]: I1123 06:56:58.130730 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f6s9r" event={"ID":"465518b5-58ac-4347-aae0-94b8f68a6b36","Type":"ContainerStarted","Data":"47892b436b2db37ec0ba8fcabba546706a985943bfb9bfa430ab2c0630963817"} Nov 23 06:56:58 crc kubenswrapper[4559]: E1123 06:56:58.132415 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f6s9r" podUID="465518b5-58ac-4347-aae0-94b8f68a6b36" Nov 23 06:56:58 crc kubenswrapper[4559]: I1123 06:56:58.133829 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-56qvk" event={"ID":"b5998049-f03f-4de4-9601-dd970e3215fe","Type":"ContainerStarted","Data":"c9f47622f25bf6cecf588287d96b48478ff3812ba4a2f813ef3a16b464b0d8ea"} Nov 23 06:56:58 crc kubenswrapper[4559]: E1123 06:56:58.136232 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\"" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-56qvk" podUID="b5998049-f03f-4de4-9601-dd970e3215fe" Nov 23 06:56:58 crc kubenswrapper[4559]: I1123 06:56:58.140520 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-gmv99" event={"ID":"ec988465-9f4e-46fe-beb3-e9032bf589d0","Type":"ContainerStarted","Data":"376a21edd20cae68ee13ab4dd3083f51977e79b775409799a0f0f40e41f3093b"} Nov 23 06:56:58 crc kubenswrapper[4559]: I1123 06:56:58.140558 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-gmv99" event={"ID":"ec988465-9f4e-46fe-beb3-e9032bf589d0","Type":"ContainerStarted","Data":"f8cbf32651f15090180ab96247659a2f91e8cf9cfa2742b78fe6b916891fc6c3"} Nov 23 06:56:58 crc kubenswrapper[4559]: I1123 06:56:58.140571 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-gmv99" event={"ID":"ec988465-9f4e-46fe-beb3-e9032bf589d0","Type":"ContainerStarted","Data":"19233b5757996599e3eee39a3aca8baa6b030f686414ecd52b4a7eebe2096d60"} Nov 23 06:56:58 crc kubenswrapper[4559]: I1123 06:56:58.140681 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-gmv99" Nov 23 06:56:58 crc kubenswrapper[4559]: I1123 06:56:58.148455 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7brkl" event={"ID":"a0f57f27-60d8-4fb8-ba76-c4ea5f3b99f6","Type":"ContainerStarted","Data":"2b7452e20288d71e3ae103736e08bb7b63d3ed8f43adaf6abdce4382d7c63042"} Nov 23 06:56:58 crc kubenswrapper[4559]: E1123 06:56:58.152385 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7brkl" podUID="a0f57f27-60d8-4fb8-ba76-c4ea5f3b99f6" Nov 23 06:56:58 crc kubenswrapper[4559]: I1123 06:56:58.209174 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-gmv99" podStartSLOduration=3.209154666 podStartE2EDuration="3.209154666s" podCreationTimestamp="2025-11-23 06:56:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:56:58.188793058 +0000 UTC m=+720.210778692" watchObservedRunningTime="2025-11-23 06:56:58.209154666 +0000 UTC m=+720.231140281" Nov 23 06:56:59 crc kubenswrapper[4559]: E1123 06:56:59.158307 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\"" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-6ht8s" podUID="38f7bf7e-6d60-4cda-a988-ef9ce22883f8" Nov 23 06:56:59 crc kubenswrapper[4559]: E1123 06:56:59.158321 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7brkl" podUID="a0f57f27-60d8-4fb8-ba76-c4ea5f3b99f6" Nov 23 06:56:59 crc kubenswrapper[4559]: E1123 06:56:59.158573 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-8hrnp" podUID="9c496b90-00da-4367-90f5-347a0c5a8ac6" Nov 23 06:56:59 crc kubenswrapper[4559]: E1123 06:56:59.158626 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c\\\"\"" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-htkt7" podUID="c5127c1b-7342-4669-b04b-8484fd2ac326" Nov 23 06:56:59 crc kubenswrapper[4559]: E1123 06:56:59.158991 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-8464cf66df-c2gjt" podUID="2463fe89-bc43-469f-9837-3ffd4b75605c" Nov 23 06:56:59 crc kubenswrapper[4559]: E1123 06:56:59.159176 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f6s9r" podUID="465518b5-58ac-4347-aae0-94b8f68a6b36" Nov 23 06:56:59 crc kubenswrapper[4559]: E1123 06:56:59.159245 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-62tqn" podUID="6486d960-766f-4622-bad1-853e2b6a3fce" Nov 23 06:56:59 crc kubenswrapper[4559]: E1123 06:56:59.159841 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\"" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-56qvk" podUID="b5998049-f03f-4de4-9601-dd970e3215fe" Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.207667 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-qgvkm" event={"ID":"50ff7c3e-8c4a-45e9-abe0-f5391a5c7c95","Type":"ContainerStarted","Data":"4405382d2fee3192abf8195068a85a05bb6331076d2be0a0821de70a820b921d"} Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.207964 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-qgvkm" event={"ID":"50ff7c3e-8c4a-45e9-abe0-f5391a5c7c95","Type":"ContainerStarted","Data":"a18d29f838a5e74a7cc39011831f4e3a465bdc4530e257a0c23de66839217110"} Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.207999 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-qgvkm" Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.212750 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-qq8sm" event={"ID":"82279b3c-bca1-4891-b7d9-a367005ad84e","Type":"ContainerStarted","Data":"f6764febb2e786a72e52bf4d2ece343fc94098af18ab5b5c3057833378abf7a0"} Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.212793 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-qq8sm" event={"ID":"82279b3c-bca1-4891-b7d9-a367005ad84e","Type":"ContainerStarted","Data":"ce1dac9b706bc71ecf7a6a09a5b7868e667fde4bbca6aea40bf1b6a188244bda"} Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.213163 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-qq8sm" Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.224714 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r" event={"ID":"14445d59-81da-4b26-bd5d-2bbe2bd1a9c0","Type":"ContainerStarted","Data":"5ff511845fe20b9f64822546e0a61dbcb6d1ae4a9c7b647acd3ea9dc1f52bce1"} Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.224743 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r" event={"ID":"14445d59-81da-4b26-bd5d-2bbe2bd1a9c0","Type":"ContainerStarted","Data":"25551f1e72ee037f94d668b62dc3fade9bfc83d81105482b8f30f2036a358ab0"} Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.225046 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r" Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.232887 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-47bnx" event={"ID":"6f79eaa5-12ee-4c47-8c5f-c0fa9190bc2d","Type":"ContainerStarted","Data":"59c90aceef8d7596cf94a790d00220b5caf410f0405f83f1dfba1a74ee202a5b"} Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.242796 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-9tvg2" event={"ID":"5b70e018-1ab3-4f23-b43f-1f433a096b81","Type":"ContainerStarted","Data":"e7def14afa8d6e4a631c6f69e1ce7e7f506e366871ae92061e6c063cf621e830"} Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.242837 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-9tvg2" event={"ID":"5b70e018-1ab3-4f23-b43f-1f433a096b81","Type":"ContainerStarted","Data":"a07bbd0e70dda6f0c35b045556a405741b5f1425646bb08c93a765aaede24ae7"} Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.243851 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-9tvg2" Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.251463 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-bg99z" event={"ID":"a659f54b-26d9-4681-963a-40348401f023","Type":"ContainerStarted","Data":"c907f38690144a79c5f1e6da38ce24c63fb40486957e6133cc4dc5b82df1269d"} Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.251996 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-bg99z" Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.258231 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-d78br" event={"ID":"35944557-7948-4d54-8225-788d51eb01d1","Type":"ContainerStarted","Data":"65845bbcfb849c86400560bfe3edbdfb13ee33b6e3849cb25acf97b554f5d2eb"} Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.258263 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-d78br" event={"ID":"35944557-7948-4d54-8225-788d51eb01d1","Type":"ContainerStarted","Data":"da6b4368453aeae400969c06bcd05377b17f2a456c3b5e5e8a0a76c6756bf33a"} Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.258571 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-d78br" Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.266929 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-ks2b7" event={"ID":"090f1d86-b4e0-43ca-9aee-49c771cab8c1","Type":"ContainerStarted","Data":"6e7f9c7cd87f86106a67671750222a6ead662fc44960de67dbe9b31349d3fc63"} Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.266958 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-ks2b7" event={"ID":"090f1d86-b4e0-43ca-9aee-49c771cab8c1","Type":"ContainerStarted","Data":"f30559333b5760251bcbfb7053e8456c1b92959b5c9c2aca184193b822a4c7c3"} Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.267117 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-ks2b7" Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.268796 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-5tcjp" event={"ID":"527a2a9e-741e-4f1b-8546-a852d73a836b","Type":"ContainerStarted","Data":"21f654932c19e68e9d19905e25ab4c2dcf0c16131cd74a6b7f36dceb6e0fa97e"} Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.280049 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-qgvkm" podStartSLOduration=2.924421992 podStartE2EDuration="10.280037268s" podCreationTimestamp="2025-11-23 06:56:55 +0000 UTC" firstStartedPulling="2025-11-23 06:56:56.659614555 +0000 UTC m=+718.681600169" lastFinishedPulling="2025-11-23 06:57:04.015229831 +0000 UTC m=+726.037215445" observedRunningTime="2025-11-23 06:57:05.233223219 +0000 UTC m=+727.255208834" watchObservedRunningTime="2025-11-23 06:57:05.280037268 +0000 UTC m=+727.302022883" Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.280578 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r" podStartSLOduration=3.2979853439999998 podStartE2EDuration="10.28057496s" podCreationTimestamp="2025-11-23 06:56:55 +0000 UTC" firstStartedPulling="2025-11-23 06:56:57.023784192 +0000 UTC m=+719.045769806" lastFinishedPulling="2025-11-23 06:57:04.006373808 +0000 UTC m=+726.028359422" observedRunningTime="2025-11-23 06:57:05.278270066 +0000 UTC m=+727.300255670" watchObservedRunningTime="2025-11-23 06:57:05.28057496 +0000 UTC m=+727.302560574" Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.283636 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-rxpxw" event={"ID":"6935f82b-929a-421a-95d0-3315c9d0ef70","Type":"ContainerStarted","Data":"65c0d17f7ad6c359d8dbfc5512293d46a00627bda952cd4b6460dd06a39c56a6"} Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.283694 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-rxpxw" event={"ID":"6935f82b-929a-421a-95d0-3315c9d0ef70","Type":"ContainerStarted","Data":"826340d67da05685ddac31b4a8e813eb7a0f453f27ecd4dfc2be79208dacd7fe"} Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.283899 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-rxpxw" Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.295237 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-hgbh8" event={"ID":"7c53dd20-9115-4fac-a49e-b6b099150245","Type":"ContainerStarted","Data":"8c36e6529acd092a09f8a0ec32d3c60c97a9352a1d91e109ce4aed1aea50d789"} Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.295269 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-hgbh8" event={"ID":"7c53dd20-9115-4fac-a49e-b6b099150245","Type":"ContainerStarted","Data":"db4e9c5a6519b828fefbadcdc4bcb81303d629508aa5e595abc88b817ddb26d7"} Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.295444 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-hgbh8" Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.300756 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-qfgkh" event={"ID":"042a59c3-a00e-4934-b16e-c731f0ab903f","Type":"ContainerStarted","Data":"faab02adf75ba39cc2fa96af306f25591a8b534f5454ae13b2457d902237522a"} Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.300785 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-qfgkh" event={"ID":"042a59c3-a00e-4934-b16e-c731f0ab903f","Type":"ContainerStarted","Data":"02869b3585bbb09409d9f9e95318dad5b1e49446bd719c50463ed0df8dbfab1b"} Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.301153 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-qfgkh" Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.303469 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-5rhv2" event={"ID":"620fe37e-07bb-4286-b1e8-fd62db1cf022","Type":"ContainerStarted","Data":"99b8ab9de38019c218d9f58146d7190388fa1869728974ca9831384036d3d84a"} Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.303494 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-5rhv2" event={"ID":"620fe37e-07bb-4286-b1e8-fd62db1cf022","Type":"ContainerStarted","Data":"41e78de6d2a513352ae8b6a468015f7ac523ab19de9903c035eb17ae5175b848"} Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.303896 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-5rhv2" Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.305831 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-c776j" event={"ID":"4cc947bb-881d-4592-9db6-ea53f6694ea5","Type":"ContainerStarted","Data":"c8bc85a2bbfa72bf7b2146f764e0d5854a32fb63161b5df2bc3b650648b94223"} Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.325657 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-rxpxw" podStartSLOduration=2.959054317 podStartE2EDuration="10.32562402s" podCreationTimestamp="2025-11-23 06:56:55 +0000 UTC" firstStartedPulling="2025-11-23 06:56:56.648095195 +0000 UTC m=+718.670080809" lastFinishedPulling="2025-11-23 06:57:04.014664898 +0000 UTC m=+726.036650512" observedRunningTime="2025-11-23 06:57:05.325147985 +0000 UTC m=+727.347133598" watchObservedRunningTime="2025-11-23 06:57:05.32562402 +0000 UTC m=+727.347609634" Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.325921 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-qq8sm" podStartSLOduration=2.76742129 podStartE2EDuration="10.325916992s" podCreationTimestamp="2025-11-23 06:56:55 +0000 UTC" firstStartedPulling="2025-11-23 06:56:56.452598402 +0000 UTC m=+718.474584016" lastFinishedPulling="2025-11-23 06:57:04.011094103 +0000 UTC m=+726.033079718" observedRunningTime="2025-11-23 06:57:05.30908475 +0000 UTC m=+727.331070364" watchObservedRunningTime="2025-11-23 06:57:05.325916992 +0000 UTC m=+727.347902605" Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.346661 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-d78br" podStartSLOduration=2.461614378 podStartE2EDuration="10.34663542s" podCreationTimestamp="2025-11-23 06:56:55 +0000 UTC" firstStartedPulling="2025-11-23 06:56:56.125657611 +0000 UTC m=+718.147643215" lastFinishedPulling="2025-11-23 06:57:04.010678642 +0000 UTC m=+726.032664257" observedRunningTime="2025-11-23 06:57:05.34291257 +0000 UTC m=+727.364898184" watchObservedRunningTime="2025-11-23 06:57:05.34663542 +0000 UTC m=+727.368621034" Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.360933 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-hgbh8" podStartSLOduration=3.145167656 podStartE2EDuration="10.36091933s" podCreationTimestamp="2025-11-23 06:56:55 +0000 UTC" firstStartedPulling="2025-11-23 06:56:56.82014608 +0000 UTC m=+718.842131694" lastFinishedPulling="2025-11-23 06:57:04.035897754 +0000 UTC m=+726.057883368" observedRunningTime="2025-11-23 06:57:05.358586443 +0000 UTC m=+727.380572057" watchObservedRunningTime="2025-11-23 06:57:05.36091933 +0000 UTC m=+727.382904944" Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.381987 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-ks2b7" podStartSLOduration=2.501570101 podStartE2EDuration="10.381965685s" podCreationTimestamp="2025-11-23 06:56:55 +0000 UTC" firstStartedPulling="2025-11-23 06:56:56.096888994 +0000 UTC m=+718.118874608" lastFinishedPulling="2025-11-23 06:57:03.977284578 +0000 UTC m=+725.999270192" observedRunningTime="2025-11-23 06:57:05.378882247 +0000 UTC m=+727.400867861" watchObservedRunningTime="2025-11-23 06:57:05.381965685 +0000 UTC m=+727.403951299" Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.393677 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-bg99z" podStartSLOduration=3.065260236 podStartE2EDuration="10.393661317s" podCreationTimestamp="2025-11-23 06:56:55 +0000 UTC" firstStartedPulling="2025-11-23 06:56:56.64622666 +0000 UTC m=+718.668212275" lastFinishedPulling="2025-11-23 06:57:03.974627742 +0000 UTC m=+725.996613356" observedRunningTime="2025-11-23 06:57:05.392393384 +0000 UTC m=+727.414378998" watchObservedRunningTime="2025-11-23 06:57:05.393661317 +0000 UTC m=+727.415646931" Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.411723 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-qfgkh" podStartSLOduration=2.498554531 podStartE2EDuration="10.411708824s" podCreationTimestamp="2025-11-23 06:56:55 +0000 UTC" firstStartedPulling="2025-11-23 06:56:56.064443855 +0000 UTC m=+718.086429469" lastFinishedPulling="2025-11-23 06:57:03.977598148 +0000 UTC m=+725.999583762" observedRunningTime="2025-11-23 06:57:05.407507203 +0000 UTC m=+727.429492818" watchObservedRunningTime="2025-11-23 06:57:05.411708824 +0000 UTC m=+727.433694438" Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.437491 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-c776j" podStartSLOduration=3.080803043 podStartE2EDuration="10.437470466s" podCreationTimestamp="2025-11-23 06:56:55 +0000 UTC" firstStartedPulling="2025-11-23 06:56:56.660402527 +0000 UTC m=+718.682388141" lastFinishedPulling="2025-11-23 06:57:04.01706995 +0000 UTC m=+726.039055564" observedRunningTime="2025-11-23 06:57:05.430428004 +0000 UTC m=+727.452413619" watchObservedRunningTime="2025-11-23 06:57:05.437470466 +0000 UTC m=+727.459456080" Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.450724 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-5rhv2" podStartSLOduration=3.076233881 podStartE2EDuration="10.45070363s" podCreationTimestamp="2025-11-23 06:56:55 +0000 UTC" firstStartedPulling="2025-11-23 06:56:56.643900497 +0000 UTC m=+718.665886111" lastFinishedPulling="2025-11-23 06:57:04.018370245 +0000 UTC m=+726.040355860" observedRunningTime="2025-11-23 06:57:05.448437349 +0000 UTC m=+727.470422963" watchObservedRunningTime="2025-11-23 06:57:05.45070363 +0000 UTC m=+727.472689234" Nov 23 06:57:05 crc kubenswrapper[4559]: I1123 06:57:05.540961 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-c776j" Nov 23 06:57:06 crc kubenswrapper[4559]: I1123 06:57:06.315151 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-bg99z" event={"ID":"a659f54b-26d9-4681-963a-40348401f023","Type":"ContainerStarted","Data":"34550649a64311b33bc91bc2888d74bbdf955ff9b3676fbda3706670a9231cb9"} Nov 23 06:57:06 crc kubenswrapper[4559]: I1123 06:57:06.317164 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-c776j" event={"ID":"4cc947bb-881d-4592-9db6-ea53f6694ea5","Type":"ContainerStarted","Data":"4d65c4739e96a034fa120a9d0d3b44262e3266d36b4a2f6534a4bd88944c4ddb"} Nov 23 06:57:06 crc kubenswrapper[4559]: I1123 06:57:06.318940 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-47bnx" event={"ID":"6f79eaa5-12ee-4c47-8c5f-c0fa9190bc2d","Type":"ContainerStarted","Data":"1c8177de9b077bf3ed227d83817bdcf4133c76530563c55c62af39c042e6de3a"} Nov 23 06:57:06 crc kubenswrapper[4559]: I1123 06:57:06.319557 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-47bnx" Nov 23 06:57:06 crc kubenswrapper[4559]: I1123 06:57:06.321781 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-5tcjp" event={"ID":"527a2a9e-741e-4f1b-8546-a852d73a836b","Type":"ContainerStarted","Data":"48e5a30e4c4d55b5e30cb9b5c9c4a94956d40c75f4e039833e76117672640478"} Nov 23 06:57:06 crc kubenswrapper[4559]: I1123 06:57:06.336806 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-47bnx" podStartSLOduration=3.391420609 podStartE2EDuration="11.336789219s" podCreationTimestamp="2025-11-23 06:56:55 +0000 UTC" firstStartedPulling="2025-11-23 06:56:56.097739374 +0000 UTC m=+718.119724987" lastFinishedPulling="2025-11-23 06:57:04.043107982 +0000 UTC m=+726.065093597" observedRunningTime="2025-11-23 06:57:06.33130136 +0000 UTC m=+728.353286973" watchObservedRunningTime="2025-11-23 06:57:06.336789219 +0000 UTC m=+728.358774833" Nov 23 06:57:06 crc kubenswrapper[4559]: I1123 06:57:06.337839 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-9tvg2" podStartSLOduration=3.778411598 podStartE2EDuration="11.337832581s" podCreationTimestamp="2025-11-23 06:56:55 +0000 UTC" firstStartedPulling="2025-11-23 06:56:56.45125271 +0000 UTC m=+718.473238325" lastFinishedPulling="2025-11-23 06:57:04.010673694 +0000 UTC m=+726.032659308" observedRunningTime="2025-11-23 06:57:05.462118094 +0000 UTC m=+727.484103708" watchObservedRunningTime="2025-11-23 06:57:06.337832581 +0000 UTC m=+728.359818195" Nov 23 06:57:06 crc kubenswrapper[4559]: I1123 06:57:06.346323 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-5tcjp" podStartSLOduration=3.982453981 podStartE2EDuration="11.346287489s" podCreationTimestamp="2025-11-23 06:56:55 +0000 UTC" firstStartedPulling="2025-11-23 06:56:56.642131189 +0000 UTC m=+718.664116804" lastFinishedPulling="2025-11-23 06:57:04.005964697 +0000 UTC m=+726.027950312" observedRunningTime="2025-11-23 06:57:06.344382947 +0000 UTC m=+728.366368571" watchObservedRunningTime="2025-11-23 06:57:06.346287489 +0000 UTC m=+728.368273103" Nov 23 06:57:06 crc kubenswrapper[4559]: I1123 06:57:06.825366 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-gmv99" Nov 23 06:57:07 crc kubenswrapper[4559]: I1123 06:57:07.329741 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-5tcjp" Nov 23 06:57:11 crc kubenswrapper[4559]: I1123 06:57:11.356089 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-8464cf66df-c2gjt" event={"ID":"2463fe89-bc43-469f-9837-3ffd4b75605c","Type":"ContainerStarted","Data":"6a8de6873b8fc37f07a6b31939bcc1ec36fad233a2bc8b1193452dfbfb4c6a59"} Nov 23 06:57:11 crc kubenswrapper[4559]: I1123 06:57:11.356839 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-8464cf66df-c2gjt" Nov 23 06:57:11 crc kubenswrapper[4559]: I1123 06:57:11.373219 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-8464cf66df-c2gjt" podStartSLOduration=2.55105376 podStartE2EDuration="16.373209241s" podCreationTimestamp="2025-11-23 06:56:55 +0000 UTC" firstStartedPulling="2025-11-23 06:56:56.820340966 +0000 UTC m=+718.842326580" lastFinishedPulling="2025-11-23 06:57:10.642496446 +0000 UTC m=+732.664482061" observedRunningTime="2025-11-23 06:57:11.371378018 +0000 UTC m=+733.393363631" watchObservedRunningTime="2025-11-23 06:57:11.373209241 +0000 UTC m=+733.395194855" Nov 23 06:57:15 crc kubenswrapper[4559]: I1123 06:57:15.417915 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-ks2b7" Nov 23 06:57:15 crc kubenswrapper[4559]: I1123 06:57:15.426445 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-d78br" Nov 23 06:57:15 crc kubenswrapper[4559]: I1123 06:57:15.470097 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-47bnx" Nov 23 06:57:15 crc kubenswrapper[4559]: I1123 06:57:15.505368 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-qfgkh" Nov 23 06:57:15 crc kubenswrapper[4559]: I1123 06:57:15.532390 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-9tvg2" Nov 23 06:57:15 crc kubenswrapper[4559]: I1123 06:57:15.546780 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-c776j" Nov 23 06:57:15 crc kubenswrapper[4559]: I1123 06:57:15.561052 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-qq8sm" Nov 23 06:57:15 crc kubenswrapper[4559]: I1123 06:57:15.569045 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-5tcjp" Nov 23 06:57:15 crc kubenswrapper[4559]: I1123 06:57:15.598277 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-rxpxw" Nov 23 06:57:15 crc kubenswrapper[4559]: I1123 06:57:15.607257 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-qgvkm" Nov 23 06:57:15 crc kubenswrapper[4559]: I1123 06:57:15.665842 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-5rhv2" Nov 23 06:57:15 crc kubenswrapper[4559]: I1123 06:57:15.670400 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-hgbh8" Nov 23 06:57:15 crc kubenswrapper[4559]: I1123 06:57:15.690813 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-bg99z" Nov 23 06:57:15 crc kubenswrapper[4559]: I1123 06:57:15.913311 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-8464cf66df-c2gjt" Nov 23 06:57:16 crc kubenswrapper[4559]: I1123 06:57:16.387631 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r" Nov 23 06:57:16 crc kubenswrapper[4559]: I1123 06:57:16.389286 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-56qvk" event={"ID":"b5998049-f03f-4de4-9601-dd970e3215fe","Type":"ContainerStarted","Data":"893b18ca387f86247b3d47a4c063f8112bf485c89bbc2b3f3a662b4df6ab66d4"} Nov 23 06:57:16 crc kubenswrapper[4559]: I1123 06:57:16.389451 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-56qvk" Nov 23 06:57:16 crc kubenswrapper[4559]: I1123 06:57:16.392190 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7brkl" event={"ID":"a0f57f27-60d8-4fb8-ba76-c4ea5f3b99f6","Type":"ContainerStarted","Data":"503d6f47dfbc75a8002ae73720c6c5f3a78a4b0d9c776887530d85de7c7f4cc1"} Nov 23 06:57:16 crc kubenswrapper[4559]: I1123 06:57:16.392388 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7brkl" Nov 23 06:57:16 crc kubenswrapper[4559]: I1123 06:57:16.393669 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-6ht8s" event={"ID":"38f7bf7e-6d60-4cda-a988-ef9ce22883f8","Type":"ContainerStarted","Data":"b6f12f95382d74f5af3ac7c3644e06ffed0b94974967666a23dd38843ba36e56"} Nov 23 06:57:16 crc kubenswrapper[4559]: I1123 06:57:16.393868 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-6ht8s" Nov 23 06:57:16 crc kubenswrapper[4559]: I1123 06:57:16.496525 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7brkl" podStartSLOduration=2.366637394 podStartE2EDuration="21.496493743s" podCreationTimestamp="2025-11-23 06:56:55 +0000 UTC" firstStartedPulling="2025-11-23 06:56:56.849591929 +0000 UTC m=+718.871577544" lastFinishedPulling="2025-11-23 06:57:15.979448279 +0000 UTC m=+738.001433893" observedRunningTime="2025-11-23 06:57:16.448866755 +0000 UTC m=+738.470852369" watchObservedRunningTime="2025-11-23 06:57:16.496493743 +0000 UTC m=+738.518479357" Nov 23 06:57:16 crc kubenswrapper[4559]: I1123 06:57:16.510273 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-56qvk" podStartSLOduration=2.383284176 podStartE2EDuration="21.510264969s" podCreationTimestamp="2025-11-23 06:56:55 +0000 UTC" firstStartedPulling="2025-11-23 06:56:56.855438413 +0000 UTC m=+718.877424027" lastFinishedPulling="2025-11-23 06:57:15.982419206 +0000 UTC m=+738.004404820" observedRunningTime="2025-11-23 06:57:16.484222418 +0000 UTC m=+738.506208032" watchObservedRunningTime="2025-11-23 06:57:16.510264969 +0000 UTC m=+738.532250583" Nov 23 06:57:16 crc kubenswrapper[4559]: I1123 06:57:16.541901 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-6ht8s" podStartSLOduration=3.947307301 podStartE2EDuration="21.541880018s" podCreationTimestamp="2025-11-23 06:56:55 +0000 UTC" firstStartedPulling="2025-11-23 06:56:56.855257062 +0000 UTC m=+718.877242676" lastFinishedPulling="2025-11-23 06:57:14.449829779 +0000 UTC m=+736.471815393" observedRunningTime="2025-11-23 06:57:16.505093094 +0000 UTC m=+738.527078708" watchObservedRunningTime="2025-11-23 06:57:16.541880018 +0000 UTC m=+738.563865621" Nov 23 06:57:17 crc kubenswrapper[4559]: I1123 06:57:17.406211 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-htkt7" event={"ID":"c5127c1b-7342-4669-b04b-8484fd2ac326","Type":"ContainerStarted","Data":"21aafb3ebf94723df1e0134ae3405288bde381afb7878553462aa8f8be329799"} Nov 23 06:57:17 crc kubenswrapper[4559]: I1123 06:57:17.407539 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-htkt7" Nov 23 06:57:17 crc kubenswrapper[4559]: I1123 06:57:17.410850 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-62tqn" event={"ID":"6486d960-766f-4622-bad1-853e2b6a3fce","Type":"ContainerStarted","Data":"71f4287a76b5a5c4be5a7447bd5977d1196d92003b0b5ec8cb8976e6c706ff80"} Nov 23 06:57:17 crc kubenswrapper[4559]: I1123 06:57:17.411295 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-62tqn" Nov 23 06:57:17 crc kubenswrapper[4559]: I1123 06:57:17.413762 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-8hrnp" event={"ID":"9c496b90-00da-4367-90f5-347a0c5a8ac6","Type":"ContainerStarted","Data":"2d09b68fa7438ab6c5cbe7b8b6314fe36680788b81210b0fc8cb6f779ec54308"} Nov 23 06:57:17 crc kubenswrapper[4559]: I1123 06:57:17.414211 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-8hrnp" Nov 23 06:57:17 crc kubenswrapper[4559]: I1123 06:57:17.416263 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f6s9r" event={"ID":"465518b5-58ac-4347-aae0-94b8f68a6b36","Type":"ContainerStarted","Data":"80d8d1fcd6e00261ad2f338af0b0897853d20b491f66144ae8c5ecd450c11e3a"} Nov 23 06:57:17 crc kubenswrapper[4559]: I1123 06:57:17.423718 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-htkt7" podStartSLOduration=3.276323811 podStartE2EDuration="22.423706697s" podCreationTimestamp="2025-11-23 06:56:55 +0000 UTC" firstStartedPulling="2025-11-23 06:56:56.835041739 +0000 UTC m=+718.857027354" lastFinishedPulling="2025-11-23 06:57:15.982424636 +0000 UTC m=+738.004410240" observedRunningTime="2025-11-23 06:57:17.420825581 +0000 UTC m=+739.442811194" watchObservedRunningTime="2025-11-23 06:57:17.423706697 +0000 UTC m=+739.445692312" Nov 23 06:57:17 crc kubenswrapper[4559]: I1123 06:57:17.435557 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-f6s9r" podStartSLOduration=3.435563166 podStartE2EDuration="22.435546462s" podCreationTimestamp="2025-11-23 06:56:55 +0000 UTC" firstStartedPulling="2025-11-23 06:56:57.024468488 +0000 UTC m=+719.046454102" lastFinishedPulling="2025-11-23 06:57:16.024451784 +0000 UTC m=+738.046437398" observedRunningTime="2025-11-23 06:57:17.433027795 +0000 UTC m=+739.455013409" watchObservedRunningTime="2025-11-23 06:57:17.435546462 +0000 UTC m=+739.457532075" Nov 23 06:57:17 crc kubenswrapper[4559]: I1123 06:57:17.443903 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-62tqn" podStartSLOduration=2.359007615 podStartE2EDuration="22.443885221s" podCreationTimestamp="2025-11-23 06:56:55 +0000 UTC" firstStartedPulling="2025-11-23 06:56:56.839091816 +0000 UTC m=+718.861077429" lastFinishedPulling="2025-11-23 06:57:16.92396942 +0000 UTC m=+738.945955035" observedRunningTime="2025-11-23 06:57:17.442734818 +0000 UTC m=+739.464720432" watchObservedRunningTime="2025-11-23 06:57:17.443885221 +0000 UTC m=+739.465870835" Nov 23 06:57:17 crc kubenswrapper[4559]: I1123 06:57:17.455779 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-8hrnp" podStartSLOduration=3.2824191320000002 podStartE2EDuration="22.455756945s" podCreationTimestamp="2025-11-23 06:56:55 +0000 UTC" firstStartedPulling="2025-11-23 06:56:56.835247296 +0000 UTC m=+718.857232910" lastFinishedPulling="2025-11-23 06:57:16.008585108 +0000 UTC m=+738.030570723" observedRunningTime="2025-11-23 06:57:17.452919439 +0000 UTC m=+739.474905053" watchObservedRunningTime="2025-11-23 06:57:17.455756945 +0000 UTC m=+739.477742558" Nov 23 06:57:18 crc kubenswrapper[4559]: I1123 06:57:18.516825 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r5ts6"] Nov 23 06:57:18 crc kubenswrapper[4559]: I1123 06:57:18.517083 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-r5ts6" podUID="534e137b-1c8d-45fb-9c1b-6804913d309b" containerName="controller-manager" containerID="cri-o://474a804b27f6b4b95c796f5b32a7a67419f708adfb40b8b4739183d0ce9df8cc" gracePeriod=30 Nov 23 06:57:18 crc kubenswrapper[4559]: I1123 06:57:18.621831 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b"] Nov 23 06:57:18 crc kubenswrapper[4559]: I1123 06:57:18.622037 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b" podUID="08b0652b-f30f-4c2c-bd1a-75b80a66b144" containerName="route-controller-manager" containerID="cri-o://742c23d558702aa7bdf98ee69c5e2808ee4470e0935d2b09dbd82585cd894eeb" gracePeriod=30 Nov 23 06:57:18 crc kubenswrapper[4559]: I1123 06:57:18.865461 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-r5ts6" Nov 23 06:57:18 crc kubenswrapper[4559]: I1123 06:57:18.928271 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/534e137b-1c8d-45fb-9c1b-6804913d309b-client-ca\") pod \"534e137b-1c8d-45fb-9c1b-6804913d309b\" (UID: \"534e137b-1c8d-45fb-9c1b-6804913d309b\") " Nov 23 06:57:18 crc kubenswrapper[4559]: I1123 06:57:18.928337 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/534e137b-1c8d-45fb-9c1b-6804913d309b-config\") pod \"534e137b-1c8d-45fb-9c1b-6804913d309b\" (UID: \"534e137b-1c8d-45fb-9c1b-6804913d309b\") " Nov 23 06:57:18 crc kubenswrapper[4559]: I1123 06:57:18.928380 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hcm8\" (UniqueName: \"kubernetes.io/projected/534e137b-1c8d-45fb-9c1b-6804913d309b-kube-api-access-8hcm8\") pod \"534e137b-1c8d-45fb-9c1b-6804913d309b\" (UID: \"534e137b-1c8d-45fb-9c1b-6804913d309b\") " Nov 23 06:57:18 crc kubenswrapper[4559]: I1123 06:57:18.928454 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/534e137b-1c8d-45fb-9c1b-6804913d309b-serving-cert\") pod \"534e137b-1c8d-45fb-9c1b-6804913d309b\" (UID: \"534e137b-1c8d-45fb-9c1b-6804913d309b\") " Nov 23 06:57:18 crc kubenswrapper[4559]: I1123 06:57:18.928489 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/534e137b-1c8d-45fb-9c1b-6804913d309b-proxy-ca-bundles\") pod \"534e137b-1c8d-45fb-9c1b-6804913d309b\" (UID: \"534e137b-1c8d-45fb-9c1b-6804913d309b\") " Nov 23 06:57:18 crc kubenswrapper[4559]: I1123 06:57:18.929482 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/534e137b-1c8d-45fb-9c1b-6804913d309b-config" (OuterVolumeSpecName: "config") pod "534e137b-1c8d-45fb-9c1b-6804913d309b" (UID: "534e137b-1c8d-45fb-9c1b-6804913d309b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:57:18 crc kubenswrapper[4559]: I1123 06:57:18.929493 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/534e137b-1c8d-45fb-9c1b-6804913d309b-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "534e137b-1c8d-45fb-9c1b-6804913d309b" (UID: "534e137b-1c8d-45fb-9c1b-6804913d309b"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:57:18 crc kubenswrapper[4559]: I1123 06:57:18.929944 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/534e137b-1c8d-45fb-9c1b-6804913d309b-client-ca" (OuterVolumeSpecName: "client-ca") pod "534e137b-1c8d-45fb-9c1b-6804913d309b" (UID: "534e137b-1c8d-45fb-9c1b-6804913d309b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:57:18 crc kubenswrapper[4559]: I1123 06:57:18.935077 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/534e137b-1c8d-45fb-9c1b-6804913d309b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "534e137b-1c8d-45fb-9c1b-6804913d309b" (UID: "534e137b-1c8d-45fb-9c1b-6804913d309b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:57:18 crc kubenswrapper[4559]: I1123 06:57:18.935092 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/534e137b-1c8d-45fb-9c1b-6804913d309b-kube-api-access-8hcm8" (OuterVolumeSpecName: "kube-api-access-8hcm8") pod "534e137b-1c8d-45fb-9c1b-6804913d309b" (UID: "534e137b-1c8d-45fb-9c1b-6804913d309b"). InnerVolumeSpecName "kube-api-access-8hcm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:57:18 crc kubenswrapper[4559]: I1123 06:57:18.959074 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.029375 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08b0652b-f30f-4c2c-bd1a-75b80a66b144-serving-cert\") pod \"08b0652b-f30f-4c2c-bd1a-75b80a66b144\" (UID: \"08b0652b-f30f-4c2c-bd1a-75b80a66b144\") " Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.029438 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08b0652b-f30f-4c2c-bd1a-75b80a66b144-config\") pod \"08b0652b-f30f-4c2c-bd1a-75b80a66b144\" (UID: \"08b0652b-f30f-4c2c-bd1a-75b80a66b144\") " Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.029465 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h642b\" (UniqueName: \"kubernetes.io/projected/08b0652b-f30f-4c2c-bd1a-75b80a66b144-kube-api-access-h642b\") pod \"08b0652b-f30f-4c2c-bd1a-75b80a66b144\" (UID: \"08b0652b-f30f-4c2c-bd1a-75b80a66b144\") " Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.029534 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/08b0652b-f30f-4c2c-bd1a-75b80a66b144-client-ca\") pod \"08b0652b-f30f-4c2c-bd1a-75b80a66b144\" (UID: \"08b0652b-f30f-4c2c-bd1a-75b80a66b144\") " Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.029744 4559 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/534e137b-1c8d-45fb-9c1b-6804913d309b-client-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.029769 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/534e137b-1c8d-45fb-9c1b-6804913d309b-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.029778 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hcm8\" (UniqueName: \"kubernetes.io/projected/534e137b-1c8d-45fb-9c1b-6804913d309b-kube-api-access-8hcm8\") on node \"crc\" DevicePath \"\"" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.029788 4559 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/534e137b-1c8d-45fb-9c1b-6804913d309b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.029796 4559 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/534e137b-1c8d-45fb-9c1b-6804913d309b-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.030351 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08b0652b-f30f-4c2c-bd1a-75b80a66b144-client-ca" (OuterVolumeSpecName: "client-ca") pod "08b0652b-f30f-4c2c-bd1a-75b80a66b144" (UID: "08b0652b-f30f-4c2c-bd1a-75b80a66b144"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.030375 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08b0652b-f30f-4c2c-bd1a-75b80a66b144-config" (OuterVolumeSpecName: "config") pod "08b0652b-f30f-4c2c-bd1a-75b80a66b144" (UID: "08b0652b-f30f-4c2c-bd1a-75b80a66b144"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.032765 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08b0652b-f30f-4c2c-bd1a-75b80a66b144-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "08b0652b-f30f-4c2c-bd1a-75b80a66b144" (UID: "08b0652b-f30f-4c2c-bd1a-75b80a66b144"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.032880 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08b0652b-f30f-4c2c-bd1a-75b80a66b144-kube-api-access-h642b" (OuterVolumeSpecName: "kube-api-access-h642b") pod "08b0652b-f30f-4c2c-bd1a-75b80a66b144" (UID: "08b0652b-f30f-4c2c-bd1a-75b80a66b144"). InnerVolumeSpecName "kube-api-access-h642b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.131370 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08b0652b-f30f-4c2c-bd1a-75b80a66b144-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.131661 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h642b\" (UniqueName: \"kubernetes.io/projected/08b0652b-f30f-4c2c-bd1a-75b80a66b144-kube-api-access-h642b\") on node \"crc\" DevicePath \"\"" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.131674 4559 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/08b0652b-f30f-4c2c-bd1a-75b80a66b144-client-ca\") on node \"crc\" DevicePath \"\"" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.131685 4559 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/08b0652b-f30f-4c2c-bd1a-75b80a66b144-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.428208 4559 generic.go:334] "Generic (PLEG): container finished" podID="08b0652b-f30f-4c2c-bd1a-75b80a66b144" containerID="742c23d558702aa7bdf98ee69c5e2808ee4470e0935d2b09dbd82585cd894eeb" exitCode=0 Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.428266 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.428284 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b" event={"ID":"08b0652b-f30f-4c2c-bd1a-75b80a66b144","Type":"ContainerDied","Data":"742c23d558702aa7bdf98ee69c5e2808ee4470e0935d2b09dbd82585cd894eeb"} Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.428315 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b" event={"ID":"08b0652b-f30f-4c2c-bd1a-75b80a66b144","Type":"ContainerDied","Data":"e65dfed27469c0d817ec33a4ee61e8bb93ff8846276827e3968f8dc3cfc014c9"} Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.428335 4559 scope.go:117] "RemoveContainer" containerID="742c23d558702aa7bdf98ee69c5e2808ee4470e0935d2b09dbd82585cd894eeb" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.429510 4559 generic.go:334] "Generic (PLEG): container finished" podID="534e137b-1c8d-45fb-9c1b-6804913d309b" containerID="474a804b27f6b4b95c796f5b32a7a67419f708adfb40b8b4739183d0ce9df8cc" exitCode=0 Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.429549 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-r5ts6" event={"ID":"534e137b-1c8d-45fb-9c1b-6804913d309b","Type":"ContainerDied","Data":"474a804b27f6b4b95c796f5b32a7a67419f708adfb40b8b4739183d0ce9df8cc"} Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.429574 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-r5ts6" event={"ID":"534e137b-1c8d-45fb-9c1b-6804913d309b","Type":"ContainerDied","Data":"72d20e034c518fb5213fbccb1e48ed36b5df5d250f4943f74f1390eef8af6220"} Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.429603 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-r5ts6" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.442575 4559 scope.go:117] "RemoveContainer" containerID="742c23d558702aa7bdf98ee69c5e2808ee4470e0935d2b09dbd82585cd894eeb" Nov 23 06:57:19 crc kubenswrapper[4559]: E1123 06:57:19.442826 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"742c23d558702aa7bdf98ee69c5e2808ee4470e0935d2b09dbd82585cd894eeb\": container with ID starting with 742c23d558702aa7bdf98ee69c5e2808ee4470e0935d2b09dbd82585cd894eeb not found: ID does not exist" containerID="742c23d558702aa7bdf98ee69c5e2808ee4470e0935d2b09dbd82585cd894eeb" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.442850 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"742c23d558702aa7bdf98ee69c5e2808ee4470e0935d2b09dbd82585cd894eeb"} err="failed to get container status \"742c23d558702aa7bdf98ee69c5e2808ee4470e0935d2b09dbd82585cd894eeb\": rpc error: code = NotFound desc = could not find container \"742c23d558702aa7bdf98ee69c5e2808ee4470e0935d2b09dbd82585cd894eeb\": container with ID starting with 742c23d558702aa7bdf98ee69c5e2808ee4470e0935d2b09dbd82585cd894eeb not found: ID does not exist" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.442929 4559 scope.go:117] "RemoveContainer" containerID="474a804b27f6b4b95c796f5b32a7a67419f708adfb40b8b4739183d0ce9df8cc" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.454691 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r5ts6"] Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.458456 4559 scope.go:117] "RemoveContainer" containerID="474a804b27f6b4b95c796f5b32a7a67419f708adfb40b8b4739183d0ce9df8cc" Nov 23 06:57:19 crc kubenswrapper[4559]: E1123 06:57:19.459034 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"474a804b27f6b4b95c796f5b32a7a67419f708adfb40b8b4739183d0ce9df8cc\": container with ID starting with 474a804b27f6b4b95c796f5b32a7a67419f708adfb40b8b4739183d0ce9df8cc not found: ID does not exist" containerID="474a804b27f6b4b95c796f5b32a7a67419f708adfb40b8b4739183d0ce9df8cc" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.459093 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"474a804b27f6b4b95c796f5b32a7a67419f708adfb40b8b4739183d0ce9df8cc"} err="failed to get container status \"474a804b27f6b4b95c796f5b32a7a67419f708adfb40b8b4739183d0ce9df8cc\": rpc error: code = NotFound desc = could not find container \"474a804b27f6b4b95c796f5b32a7a67419f708adfb40b8b4739183d0ce9df8cc\": container with ID starting with 474a804b27f6b4b95c796f5b32a7a67419f708adfb40b8b4739183d0ce9df8cc not found: ID does not exist" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.467405 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r5ts6"] Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.475716 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b"] Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.479355 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzh5b"] Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.848963 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6dc5cf565b-27kvl"] Nov 23 06:57:19 crc kubenswrapper[4559]: E1123 06:57:19.849297 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08b0652b-f30f-4c2c-bd1a-75b80a66b144" containerName="route-controller-manager" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.849310 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="08b0652b-f30f-4c2c-bd1a-75b80a66b144" containerName="route-controller-manager" Nov 23 06:57:19 crc kubenswrapper[4559]: E1123 06:57:19.849341 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="534e137b-1c8d-45fb-9c1b-6804913d309b" containerName="controller-manager" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.849347 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="534e137b-1c8d-45fb-9c1b-6804913d309b" containerName="controller-manager" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.849491 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="08b0652b-f30f-4c2c-bd1a-75b80a66b144" containerName="route-controller-manager" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.849503 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="534e137b-1c8d-45fb-9c1b-6804913d309b" containerName="controller-manager" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.849994 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6dc5cf565b-27kvl" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.852038 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.852203 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.852252 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.852534 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.852599 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.852855 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.853439 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-795677c985-4pdb8"] Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.854355 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-795677c985-4pdb8" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.856180 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.856203 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.856288 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.856503 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.857001 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.860238 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.862541 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.863974 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6dc5cf565b-27kvl"] Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.874876 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-795677c985-4pdb8"] Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.943117 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4xtl\" (UniqueName: \"kubernetes.io/projected/6fa7bc3c-ff0f-4134-a77e-0dae6acc4d53-kube-api-access-v4xtl\") pod \"route-controller-manager-6dc5cf565b-27kvl\" (UID: \"6fa7bc3c-ff0f-4134-a77e-0dae6acc4d53\") " pod="openshift-route-controller-manager/route-controller-manager-6dc5cf565b-27kvl" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.943170 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b982d29a-bb05-4f3d-b804-85c1887070a7-config\") pod \"controller-manager-795677c985-4pdb8\" (UID: \"b982d29a-bb05-4f3d-b804-85c1887070a7\") " pod="openshift-controller-manager/controller-manager-795677c985-4pdb8" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.943269 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b982d29a-bb05-4f3d-b804-85c1887070a7-proxy-ca-bundles\") pod \"controller-manager-795677c985-4pdb8\" (UID: \"b982d29a-bb05-4f3d-b804-85c1887070a7\") " pod="openshift-controller-manager/controller-manager-795677c985-4pdb8" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.943309 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x285g\" (UniqueName: \"kubernetes.io/projected/b982d29a-bb05-4f3d-b804-85c1887070a7-kube-api-access-x285g\") pod \"controller-manager-795677c985-4pdb8\" (UID: \"b982d29a-bb05-4f3d-b804-85c1887070a7\") " pod="openshift-controller-manager/controller-manager-795677c985-4pdb8" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.943344 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fa7bc3c-ff0f-4134-a77e-0dae6acc4d53-config\") pod \"route-controller-manager-6dc5cf565b-27kvl\" (UID: \"6fa7bc3c-ff0f-4134-a77e-0dae6acc4d53\") " pod="openshift-route-controller-manager/route-controller-manager-6dc5cf565b-27kvl" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.943540 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6fa7bc3c-ff0f-4134-a77e-0dae6acc4d53-serving-cert\") pod \"route-controller-manager-6dc5cf565b-27kvl\" (UID: \"6fa7bc3c-ff0f-4134-a77e-0dae6acc4d53\") " pod="openshift-route-controller-manager/route-controller-manager-6dc5cf565b-27kvl" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.943687 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b982d29a-bb05-4f3d-b804-85c1887070a7-serving-cert\") pod \"controller-manager-795677c985-4pdb8\" (UID: \"b982d29a-bb05-4f3d-b804-85c1887070a7\") " pod="openshift-controller-manager/controller-manager-795677c985-4pdb8" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.943751 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b982d29a-bb05-4f3d-b804-85c1887070a7-client-ca\") pod \"controller-manager-795677c985-4pdb8\" (UID: \"b982d29a-bb05-4f3d-b804-85c1887070a7\") " pod="openshift-controller-manager/controller-manager-795677c985-4pdb8" Nov 23 06:57:19 crc kubenswrapper[4559]: I1123 06:57:19.943831 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6fa7bc3c-ff0f-4134-a77e-0dae6acc4d53-client-ca\") pod \"route-controller-manager-6dc5cf565b-27kvl\" (UID: \"6fa7bc3c-ff0f-4134-a77e-0dae6acc4d53\") " pod="openshift-route-controller-manager/route-controller-manager-6dc5cf565b-27kvl" Nov 23 06:57:20 crc kubenswrapper[4559]: I1123 06:57:20.045353 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b982d29a-bb05-4f3d-b804-85c1887070a7-serving-cert\") pod \"controller-manager-795677c985-4pdb8\" (UID: \"b982d29a-bb05-4f3d-b804-85c1887070a7\") " pod="openshift-controller-manager/controller-manager-795677c985-4pdb8" Nov 23 06:57:20 crc kubenswrapper[4559]: I1123 06:57:20.045400 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b982d29a-bb05-4f3d-b804-85c1887070a7-client-ca\") pod \"controller-manager-795677c985-4pdb8\" (UID: \"b982d29a-bb05-4f3d-b804-85c1887070a7\") " pod="openshift-controller-manager/controller-manager-795677c985-4pdb8" Nov 23 06:57:20 crc kubenswrapper[4559]: I1123 06:57:20.045432 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6fa7bc3c-ff0f-4134-a77e-0dae6acc4d53-client-ca\") pod \"route-controller-manager-6dc5cf565b-27kvl\" (UID: \"6fa7bc3c-ff0f-4134-a77e-0dae6acc4d53\") " pod="openshift-route-controller-manager/route-controller-manager-6dc5cf565b-27kvl" Nov 23 06:57:20 crc kubenswrapper[4559]: I1123 06:57:20.045454 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4xtl\" (UniqueName: \"kubernetes.io/projected/6fa7bc3c-ff0f-4134-a77e-0dae6acc4d53-kube-api-access-v4xtl\") pod \"route-controller-manager-6dc5cf565b-27kvl\" (UID: \"6fa7bc3c-ff0f-4134-a77e-0dae6acc4d53\") " pod="openshift-route-controller-manager/route-controller-manager-6dc5cf565b-27kvl" Nov 23 06:57:20 crc kubenswrapper[4559]: I1123 06:57:20.045472 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b982d29a-bb05-4f3d-b804-85c1887070a7-config\") pod \"controller-manager-795677c985-4pdb8\" (UID: \"b982d29a-bb05-4f3d-b804-85c1887070a7\") " pod="openshift-controller-manager/controller-manager-795677c985-4pdb8" Nov 23 06:57:20 crc kubenswrapper[4559]: I1123 06:57:20.045495 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b982d29a-bb05-4f3d-b804-85c1887070a7-proxy-ca-bundles\") pod \"controller-manager-795677c985-4pdb8\" (UID: \"b982d29a-bb05-4f3d-b804-85c1887070a7\") " pod="openshift-controller-manager/controller-manager-795677c985-4pdb8" Nov 23 06:57:20 crc kubenswrapper[4559]: I1123 06:57:20.045516 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x285g\" (UniqueName: \"kubernetes.io/projected/b982d29a-bb05-4f3d-b804-85c1887070a7-kube-api-access-x285g\") pod \"controller-manager-795677c985-4pdb8\" (UID: \"b982d29a-bb05-4f3d-b804-85c1887070a7\") " pod="openshift-controller-manager/controller-manager-795677c985-4pdb8" Nov 23 06:57:20 crc kubenswrapper[4559]: I1123 06:57:20.045537 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fa7bc3c-ff0f-4134-a77e-0dae6acc4d53-config\") pod \"route-controller-manager-6dc5cf565b-27kvl\" (UID: \"6fa7bc3c-ff0f-4134-a77e-0dae6acc4d53\") " pod="openshift-route-controller-manager/route-controller-manager-6dc5cf565b-27kvl" Nov 23 06:57:20 crc kubenswrapper[4559]: I1123 06:57:20.045572 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6fa7bc3c-ff0f-4134-a77e-0dae6acc4d53-serving-cert\") pod \"route-controller-manager-6dc5cf565b-27kvl\" (UID: \"6fa7bc3c-ff0f-4134-a77e-0dae6acc4d53\") " pod="openshift-route-controller-manager/route-controller-manager-6dc5cf565b-27kvl" Nov 23 06:57:20 crc kubenswrapper[4559]: I1123 06:57:20.047192 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6fa7bc3c-ff0f-4134-a77e-0dae6acc4d53-client-ca\") pod \"route-controller-manager-6dc5cf565b-27kvl\" (UID: \"6fa7bc3c-ff0f-4134-a77e-0dae6acc4d53\") " pod="openshift-route-controller-manager/route-controller-manager-6dc5cf565b-27kvl" Nov 23 06:57:20 crc kubenswrapper[4559]: I1123 06:57:20.047222 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b982d29a-bb05-4f3d-b804-85c1887070a7-client-ca\") pod \"controller-manager-795677c985-4pdb8\" (UID: \"b982d29a-bb05-4f3d-b804-85c1887070a7\") " pod="openshift-controller-manager/controller-manager-795677c985-4pdb8" Nov 23 06:57:20 crc kubenswrapper[4559]: I1123 06:57:20.047279 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b982d29a-bb05-4f3d-b804-85c1887070a7-proxy-ca-bundles\") pod \"controller-manager-795677c985-4pdb8\" (UID: \"b982d29a-bb05-4f3d-b804-85c1887070a7\") " pod="openshift-controller-manager/controller-manager-795677c985-4pdb8" Nov 23 06:57:20 crc kubenswrapper[4559]: I1123 06:57:20.047548 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b982d29a-bb05-4f3d-b804-85c1887070a7-config\") pod \"controller-manager-795677c985-4pdb8\" (UID: \"b982d29a-bb05-4f3d-b804-85c1887070a7\") " pod="openshift-controller-manager/controller-manager-795677c985-4pdb8" Nov 23 06:57:20 crc kubenswrapper[4559]: I1123 06:57:20.047713 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fa7bc3c-ff0f-4134-a77e-0dae6acc4d53-config\") pod \"route-controller-manager-6dc5cf565b-27kvl\" (UID: \"6fa7bc3c-ff0f-4134-a77e-0dae6acc4d53\") " pod="openshift-route-controller-manager/route-controller-manager-6dc5cf565b-27kvl" Nov 23 06:57:20 crc kubenswrapper[4559]: I1123 06:57:20.049956 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6fa7bc3c-ff0f-4134-a77e-0dae6acc4d53-serving-cert\") pod \"route-controller-manager-6dc5cf565b-27kvl\" (UID: \"6fa7bc3c-ff0f-4134-a77e-0dae6acc4d53\") " pod="openshift-route-controller-manager/route-controller-manager-6dc5cf565b-27kvl" Nov 23 06:57:20 crc kubenswrapper[4559]: I1123 06:57:20.050152 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b982d29a-bb05-4f3d-b804-85c1887070a7-serving-cert\") pod \"controller-manager-795677c985-4pdb8\" (UID: \"b982d29a-bb05-4f3d-b804-85c1887070a7\") " pod="openshift-controller-manager/controller-manager-795677c985-4pdb8" Nov 23 06:57:20 crc kubenswrapper[4559]: I1123 06:57:20.061024 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x285g\" (UniqueName: \"kubernetes.io/projected/b982d29a-bb05-4f3d-b804-85c1887070a7-kube-api-access-x285g\") pod \"controller-manager-795677c985-4pdb8\" (UID: \"b982d29a-bb05-4f3d-b804-85c1887070a7\") " pod="openshift-controller-manager/controller-manager-795677c985-4pdb8" Nov 23 06:57:20 crc kubenswrapper[4559]: I1123 06:57:20.061433 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4xtl\" (UniqueName: \"kubernetes.io/projected/6fa7bc3c-ff0f-4134-a77e-0dae6acc4d53-kube-api-access-v4xtl\") pod \"route-controller-manager-6dc5cf565b-27kvl\" (UID: \"6fa7bc3c-ff0f-4134-a77e-0dae6acc4d53\") " pod="openshift-route-controller-manager/route-controller-manager-6dc5cf565b-27kvl" Nov 23 06:57:20 crc kubenswrapper[4559]: I1123 06:57:20.164791 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6dc5cf565b-27kvl" Nov 23 06:57:20 crc kubenswrapper[4559]: I1123 06:57:20.176891 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-795677c985-4pdb8" Nov 23 06:57:20 crc kubenswrapper[4559]: I1123 06:57:20.283248 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08b0652b-f30f-4c2c-bd1a-75b80a66b144" path="/var/lib/kubelet/pods/08b0652b-f30f-4c2c-bd1a-75b80a66b144/volumes" Nov 23 06:57:20 crc kubenswrapper[4559]: I1123 06:57:20.284038 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="534e137b-1c8d-45fb-9c1b-6804913d309b" path="/var/lib/kubelet/pods/534e137b-1c8d-45fb-9c1b-6804913d309b/volumes" Nov 23 06:57:20 crc kubenswrapper[4559]: I1123 06:57:20.565593 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6dc5cf565b-27kvl"] Nov 23 06:57:20 crc kubenswrapper[4559]: I1123 06:57:20.570841 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-795677c985-4pdb8"] Nov 23 06:57:20 crc kubenswrapper[4559]: W1123 06:57:20.574079 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb982d29a_bb05_4f3d_b804_85c1887070a7.slice/crio-ad7e6478e2488210716fb70e30463d7420cfeeadc0f7c667db783318fdd25331 WatchSource:0}: Error finding container ad7e6478e2488210716fb70e30463d7420cfeeadc0f7c667db783318fdd25331: Status 404 returned error can't find the container with id ad7e6478e2488210716fb70e30463d7420cfeeadc0f7c667db783318fdd25331 Nov 23 06:57:20 crc kubenswrapper[4559]: W1123 06:57:20.576024 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6fa7bc3c_ff0f_4134_a77e_0dae6acc4d53.slice/crio-9b7b9bfe1c95102a122f23dc4409430dc074c72d25ca49ef34dbbbb24a5e6721 WatchSource:0}: Error finding container 9b7b9bfe1c95102a122f23dc4409430dc074c72d25ca49ef34dbbbb24a5e6721: Status 404 returned error can't find the container with id 9b7b9bfe1c95102a122f23dc4409430dc074c72d25ca49ef34dbbbb24a5e6721 Nov 23 06:57:21 crc kubenswrapper[4559]: I1123 06:57:21.444430 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6dc5cf565b-27kvl" event={"ID":"6fa7bc3c-ff0f-4134-a77e-0dae6acc4d53","Type":"ContainerStarted","Data":"769764de5f913091f8eec4c03255b0a0d43664609fd124bfa40b5ab0264b89bc"} Nov 23 06:57:21 crc kubenswrapper[4559]: I1123 06:57:21.444940 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6dc5cf565b-27kvl" event={"ID":"6fa7bc3c-ff0f-4134-a77e-0dae6acc4d53","Type":"ContainerStarted","Data":"9b7b9bfe1c95102a122f23dc4409430dc074c72d25ca49ef34dbbbb24a5e6721"} Nov 23 06:57:21 crc kubenswrapper[4559]: I1123 06:57:21.444980 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6dc5cf565b-27kvl" Nov 23 06:57:21 crc kubenswrapper[4559]: I1123 06:57:21.447095 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-795677c985-4pdb8" event={"ID":"b982d29a-bb05-4f3d-b804-85c1887070a7","Type":"ContainerStarted","Data":"63a35352cd1c1199aadc6255315fad786354adf11186a8d602a6428f5202d7cf"} Nov 23 06:57:21 crc kubenswrapper[4559]: I1123 06:57:21.447143 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-795677c985-4pdb8" event={"ID":"b982d29a-bb05-4f3d-b804-85c1887070a7","Type":"ContainerStarted","Data":"ad7e6478e2488210716fb70e30463d7420cfeeadc0f7c667db783318fdd25331"} Nov 23 06:57:21 crc kubenswrapper[4559]: I1123 06:57:21.447338 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-795677c985-4pdb8" Nov 23 06:57:21 crc kubenswrapper[4559]: I1123 06:57:21.450378 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6dc5cf565b-27kvl" Nov 23 06:57:21 crc kubenswrapper[4559]: I1123 06:57:21.452125 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-795677c985-4pdb8" Nov 23 06:57:21 crc kubenswrapper[4559]: I1123 06:57:21.466668 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6dc5cf565b-27kvl" podStartSLOduration=3.466636046 podStartE2EDuration="3.466636046s" podCreationTimestamp="2025-11-23 06:57:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:57:21.459071442 +0000 UTC m=+743.481057057" watchObservedRunningTime="2025-11-23 06:57:21.466636046 +0000 UTC m=+743.488621660" Nov 23 06:57:21 crc kubenswrapper[4559]: I1123 06:57:21.475843 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-795677c985-4pdb8" podStartSLOduration=3.475820187 podStartE2EDuration="3.475820187s" podCreationTimestamp="2025-11-23 06:57:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:57:21.47301413 +0000 UTC m=+743.494999744" watchObservedRunningTime="2025-11-23 06:57:21.475820187 +0000 UTC m=+743.497805801" Nov 23 06:57:23 crc kubenswrapper[4559]: I1123 06:57:23.773618 4559 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 23 06:57:25 crc kubenswrapper[4559]: I1123 06:57:25.683686 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-6ht8s" Nov 23 06:57:25 crc kubenswrapper[4559]: I1123 06:57:25.712460 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-8hrnp" Nov 23 06:57:25 crc kubenswrapper[4559]: I1123 06:57:25.783513 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-htkt7" Nov 23 06:57:25 crc kubenswrapper[4559]: I1123 06:57:25.808784 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-56qvk" Nov 23 06:57:25 crc kubenswrapper[4559]: I1123 06:57:25.851614 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-62tqn" Nov 23 06:57:26 crc kubenswrapper[4559]: I1123 06:57:26.066250 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7brkl" Nov 23 06:57:26 crc kubenswrapper[4559]: I1123 06:57:26.167225 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 06:57:26 crc kubenswrapper[4559]: I1123 06:57:26.167288 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 06:57:26 crc kubenswrapper[4559]: I1123 06:57:26.167340 4559 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" Nov 23 06:57:26 crc kubenswrapper[4559]: I1123 06:57:26.168216 4559 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c1a47810490d41ffdaf229d9abf61dc047a9840ab17302f7f8ed9bcafa8fe6db"} pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 06:57:26 crc kubenswrapper[4559]: I1123 06:57:26.168280 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" containerID="cri-o://c1a47810490d41ffdaf229d9abf61dc047a9840ab17302f7f8ed9bcafa8fe6db" gracePeriod=600 Nov 23 06:57:26 crc kubenswrapper[4559]: I1123 06:57:26.475339 4559 generic.go:334] "Generic (PLEG): container finished" podID="4731beee-0cac-4189-8a70-743b0b709095" containerID="c1a47810490d41ffdaf229d9abf61dc047a9840ab17302f7f8ed9bcafa8fe6db" exitCode=0 Nov 23 06:57:26 crc kubenswrapper[4559]: I1123 06:57:26.475409 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerDied","Data":"c1a47810490d41ffdaf229d9abf61dc047a9840ab17302f7f8ed9bcafa8fe6db"} Nov 23 06:57:26 crc kubenswrapper[4559]: I1123 06:57:26.475656 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerStarted","Data":"667a247ddea953ad2f29e458b8a21398c7c8c20545d619492b9346eff264de8f"} Nov 23 06:57:26 crc kubenswrapper[4559]: I1123 06:57:26.475676 4559 scope.go:117] "RemoveContainer" containerID="caa8c42bf6f598aae815f53f4281cd35582267034e69906ef9d6b656540ea6ab" Nov 23 06:57:39 crc kubenswrapper[4559]: I1123 06:57:39.724196 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6794664cc7-vmnd9"] Nov 23 06:57:39 crc kubenswrapper[4559]: I1123 06:57:39.726483 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6794664cc7-vmnd9" Nov 23 06:57:39 crc kubenswrapper[4559]: I1123 06:57:39.729078 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 23 06:57:39 crc kubenswrapper[4559]: I1123 06:57:39.729107 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 23 06:57:39 crc kubenswrapper[4559]: I1123 06:57:39.729192 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-wqzz9" Nov 23 06:57:39 crc kubenswrapper[4559]: I1123 06:57:39.732517 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 23 06:57:39 crc kubenswrapper[4559]: I1123 06:57:39.733008 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6794664cc7-vmnd9"] Nov 23 06:57:39 crc kubenswrapper[4559]: I1123 06:57:39.772371 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84bd59c769-rdgpv"] Nov 23 06:57:39 crc kubenswrapper[4559]: I1123 06:57:39.773721 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84bd59c769-rdgpv" Nov 23 06:57:39 crc kubenswrapper[4559]: I1123 06:57:39.775335 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 23 06:57:39 crc kubenswrapper[4559]: I1123 06:57:39.779764 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84bd59c769-rdgpv"] Nov 23 06:57:39 crc kubenswrapper[4559]: I1123 06:57:39.898735 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9caa7f7e-01ab-4998-82bd-df9380117fb9-config\") pod \"dnsmasq-dns-6794664cc7-vmnd9\" (UID: \"9caa7f7e-01ab-4998-82bd-df9380117fb9\") " pod="openstack/dnsmasq-dns-6794664cc7-vmnd9" Nov 23 06:57:39 crc kubenswrapper[4559]: I1123 06:57:39.898774 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q64zk\" (UniqueName: \"kubernetes.io/projected/9caa7f7e-01ab-4998-82bd-df9380117fb9-kube-api-access-q64zk\") pod \"dnsmasq-dns-6794664cc7-vmnd9\" (UID: \"9caa7f7e-01ab-4998-82bd-df9380117fb9\") " pod="openstack/dnsmasq-dns-6794664cc7-vmnd9" Nov 23 06:57:39 crc kubenswrapper[4559]: I1123 06:57:39.898995 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2zrd\" (UniqueName: \"kubernetes.io/projected/19172889-a652-4d9e-a074-1dc45485752f-kube-api-access-d2zrd\") pod \"dnsmasq-dns-84bd59c769-rdgpv\" (UID: \"19172889-a652-4d9e-a074-1dc45485752f\") " pod="openstack/dnsmasq-dns-84bd59c769-rdgpv" Nov 23 06:57:39 crc kubenswrapper[4559]: I1123 06:57:39.899087 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19172889-a652-4d9e-a074-1dc45485752f-config\") pod \"dnsmasq-dns-84bd59c769-rdgpv\" (UID: \"19172889-a652-4d9e-a074-1dc45485752f\") " pod="openstack/dnsmasq-dns-84bd59c769-rdgpv" Nov 23 06:57:39 crc kubenswrapper[4559]: I1123 06:57:39.899114 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19172889-a652-4d9e-a074-1dc45485752f-dns-svc\") pod \"dnsmasq-dns-84bd59c769-rdgpv\" (UID: \"19172889-a652-4d9e-a074-1dc45485752f\") " pod="openstack/dnsmasq-dns-84bd59c769-rdgpv" Nov 23 06:57:40 crc kubenswrapper[4559]: I1123 06:57:40.000075 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2zrd\" (UniqueName: \"kubernetes.io/projected/19172889-a652-4d9e-a074-1dc45485752f-kube-api-access-d2zrd\") pod \"dnsmasq-dns-84bd59c769-rdgpv\" (UID: \"19172889-a652-4d9e-a074-1dc45485752f\") " pod="openstack/dnsmasq-dns-84bd59c769-rdgpv" Nov 23 06:57:40 crc kubenswrapper[4559]: I1123 06:57:40.000137 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19172889-a652-4d9e-a074-1dc45485752f-config\") pod \"dnsmasq-dns-84bd59c769-rdgpv\" (UID: \"19172889-a652-4d9e-a074-1dc45485752f\") " pod="openstack/dnsmasq-dns-84bd59c769-rdgpv" Nov 23 06:57:40 crc kubenswrapper[4559]: I1123 06:57:40.000159 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19172889-a652-4d9e-a074-1dc45485752f-dns-svc\") pod \"dnsmasq-dns-84bd59c769-rdgpv\" (UID: \"19172889-a652-4d9e-a074-1dc45485752f\") " pod="openstack/dnsmasq-dns-84bd59c769-rdgpv" Nov 23 06:57:40 crc kubenswrapper[4559]: I1123 06:57:40.000195 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9caa7f7e-01ab-4998-82bd-df9380117fb9-config\") pod \"dnsmasq-dns-6794664cc7-vmnd9\" (UID: \"9caa7f7e-01ab-4998-82bd-df9380117fb9\") " pod="openstack/dnsmasq-dns-6794664cc7-vmnd9" Nov 23 06:57:40 crc kubenswrapper[4559]: I1123 06:57:40.000213 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q64zk\" (UniqueName: \"kubernetes.io/projected/9caa7f7e-01ab-4998-82bd-df9380117fb9-kube-api-access-q64zk\") pod \"dnsmasq-dns-6794664cc7-vmnd9\" (UID: \"9caa7f7e-01ab-4998-82bd-df9380117fb9\") " pod="openstack/dnsmasq-dns-6794664cc7-vmnd9" Nov 23 06:57:40 crc kubenswrapper[4559]: I1123 06:57:40.001283 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9caa7f7e-01ab-4998-82bd-df9380117fb9-config\") pod \"dnsmasq-dns-6794664cc7-vmnd9\" (UID: \"9caa7f7e-01ab-4998-82bd-df9380117fb9\") " pod="openstack/dnsmasq-dns-6794664cc7-vmnd9" Nov 23 06:57:40 crc kubenswrapper[4559]: I1123 06:57:40.001298 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19172889-a652-4d9e-a074-1dc45485752f-dns-svc\") pod \"dnsmasq-dns-84bd59c769-rdgpv\" (UID: \"19172889-a652-4d9e-a074-1dc45485752f\") " pod="openstack/dnsmasq-dns-84bd59c769-rdgpv" Nov 23 06:57:40 crc kubenswrapper[4559]: I1123 06:57:40.001285 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19172889-a652-4d9e-a074-1dc45485752f-config\") pod \"dnsmasq-dns-84bd59c769-rdgpv\" (UID: \"19172889-a652-4d9e-a074-1dc45485752f\") " pod="openstack/dnsmasq-dns-84bd59c769-rdgpv" Nov 23 06:57:40 crc kubenswrapper[4559]: I1123 06:57:40.015943 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2zrd\" (UniqueName: \"kubernetes.io/projected/19172889-a652-4d9e-a074-1dc45485752f-kube-api-access-d2zrd\") pod \"dnsmasq-dns-84bd59c769-rdgpv\" (UID: \"19172889-a652-4d9e-a074-1dc45485752f\") " pod="openstack/dnsmasq-dns-84bd59c769-rdgpv" Nov 23 06:57:40 crc kubenswrapper[4559]: I1123 06:57:40.015954 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q64zk\" (UniqueName: \"kubernetes.io/projected/9caa7f7e-01ab-4998-82bd-df9380117fb9-kube-api-access-q64zk\") pod \"dnsmasq-dns-6794664cc7-vmnd9\" (UID: \"9caa7f7e-01ab-4998-82bd-df9380117fb9\") " pod="openstack/dnsmasq-dns-6794664cc7-vmnd9" Nov 23 06:57:40 crc kubenswrapper[4559]: I1123 06:57:40.042421 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6794664cc7-vmnd9" Nov 23 06:57:40 crc kubenswrapper[4559]: I1123 06:57:40.087696 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84bd59c769-rdgpv" Nov 23 06:57:40 crc kubenswrapper[4559]: I1123 06:57:40.431316 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6794664cc7-vmnd9"] Nov 23 06:57:40 crc kubenswrapper[4559]: W1123 06:57:40.434541 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9caa7f7e_01ab_4998_82bd_df9380117fb9.slice/crio-47bb93c7efe950640763715beffa55bf04af109f5eb4d8447badc01b6670cc24 WatchSource:0}: Error finding container 47bb93c7efe950640763715beffa55bf04af109f5eb4d8447badc01b6670cc24: Status 404 returned error can't find the container with id 47bb93c7efe950640763715beffa55bf04af109f5eb4d8447badc01b6670cc24 Nov 23 06:57:40 crc kubenswrapper[4559]: I1123 06:57:40.518630 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84bd59c769-rdgpv"] Nov 23 06:57:40 crc kubenswrapper[4559]: W1123 06:57:40.520407 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19172889_a652_4d9e_a074_1dc45485752f.slice/crio-05137c958571eb560e0bc8eb13fcb49afbce22fe98cf86ed0784f9f868a8a21e WatchSource:0}: Error finding container 05137c958571eb560e0bc8eb13fcb49afbce22fe98cf86ed0784f9f868a8a21e: Status 404 returned error can't find the container with id 05137c958571eb560e0bc8eb13fcb49afbce22fe98cf86ed0784f9f868a8a21e Nov 23 06:57:40 crc kubenswrapper[4559]: I1123 06:57:40.576523 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84bd59c769-rdgpv" event={"ID":"19172889-a652-4d9e-a074-1dc45485752f","Type":"ContainerStarted","Data":"05137c958571eb560e0bc8eb13fcb49afbce22fe98cf86ed0784f9f868a8a21e"} Nov 23 06:57:40 crc kubenswrapper[4559]: I1123 06:57:40.577513 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6794664cc7-vmnd9" event={"ID":"9caa7f7e-01ab-4998-82bd-df9380117fb9","Type":"ContainerStarted","Data":"47bb93c7efe950640763715beffa55bf04af109f5eb4d8447badc01b6670cc24"} Nov 23 06:57:42 crc kubenswrapper[4559]: I1123 06:57:42.810259 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6794664cc7-vmnd9"] Nov 23 06:57:42 crc kubenswrapper[4559]: I1123 06:57:42.837229 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f47fdfb89-xjlhn"] Nov 23 06:57:42 crc kubenswrapper[4559]: I1123 06:57:42.849058 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f47fdfb89-xjlhn" Nov 23 06:57:42 crc kubenswrapper[4559]: I1123 06:57:42.856234 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f47fdfb89-xjlhn"] Nov 23 06:57:42 crc kubenswrapper[4559]: I1123 06:57:42.939238 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebba232a-99bb-4f45-a0ce-3628f42dcc77-config\") pod \"dnsmasq-dns-7f47fdfb89-xjlhn\" (UID: \"ebba232a-99bb-4f45-a0ce-3628f42dcc77\") " pod="openstack/dnsmasq-dns-7f47fdfb89-xjlhn" Nov 23 06:57:42 crc kubenswrapper[4559]: I1123 06:57:42.939723 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ebba232a-99bb-4f45-a0ce-3628f42dcc77-dns-svc\") pod \"dnsmasq-dns-7f47fdfb89-xjlhn\" (UID: \"ebba232a-99bb-4f45-a0ce-3628f42dcc77\") " pod="openstack/dnsmasq-dns-7f47fdfb89-xjlhn" Nov 23 06:57:42 crc kubenswrapper[4559]: I1123 06:57:42.939784 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m62j4\" (UniqueName: \"kubernetes.io/projected/ebba232a-99bb-4f45-a0ce-3628f42dcc77-kube-api-access-m62j4\") pod \"dnsmasq-dns-7f47fdfb89-xjlhn\" (UID: \"ebba232a-99bb-4f45-a0ce-3628f42dcc77\") " pod="openstack/dnsmasq-dns-7f47fdfb89-xjlhn" Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.041294 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ebba232a-99bb-4f45-a0ce-3628f42dcc77-dns-svc\") pod \"dnsmasq-dns-7f47fdfb89-xjlhn\" (UID: \"ebba232a-99bb-4f45-a0ce-3628f42dcc77\") " pod="openstack/dnsmasq-dns-7f47fdfb89-xjlhn" Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.041371 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m62j4\" (UniqueName: \"kubernetes.io/projected/ebba232a-99bb-4f45-a0ce-3628f42dcc77-kube-api-access-m62j4\") pod \"dnsmasq-dns-7f47fdfb89-xjlhn\" (UID: \"ebba232a-99bb-4f45-a0ce-3628f42dcc77\") " pod="openstack/dnsmasq-dns-7f47fdfb89-xjlhn" Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.041441 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebba232a-99bb-4f45-a0ce-3628f42dcc77-config\") pod \"dnsmasq-dns-7f47fdfb89-xjlhn\" (UID: \"ebba232a-99bb-4f45-a0ce-3628f42dcc77\") " pod="openstack/dnsmasq-dns-7f47fdfb89-xjlhn" Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.042274 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebba232a-99bb-4f45-a0ce-3628f42dcc77-config\") pod \"dnsmasq-dns-7f47fdfb89-xjlhn\" (UID: \"ebba232a-99bb-4f45-a0ce-3628f42dcc77\") " pod="openstack/dnsmasq-dns-7f47fdfb89-xjlhn" Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.042816 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ebba232a-99bb-4f45-a0ce-3628f42dcc77-dns-svc\") pod \"dnsmasq-dns-7f47fdfb89-xjlhn\" (UID: \"ebba232a-99bb-4f45-a0ce-3628f42dcc77\") " pod="openstack/dnsmasq-dns-7f47fdfb89-xjlhn" Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.086918 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m62j4\" (UniqueName: \"kubernetes.io/projected/ebba232a-99bb-4f45-a0ce-3628f42dcc77-kube-api-access-m62j4\") pod \"dnsmasq-dns-7f47fdfb89-xjlhn\" (UID: \"ebba232a-99bb-4f45-a0ce-3628f42dcc77\") " pod="openstack/dnsmasq-dns-7f47fdfb89-xjlhn" Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.095630 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84bd59c769-rdgpv"] Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.121517 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-759c6cc4df-s47tx"] Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.123715 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-759c6cc4df-s47tx" Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.134034 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-759c6cc4df-s47tx"] Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.174959 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f47fdfb89-xjlhn" Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.252099 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6jq2\" (UniqueName: \"kubernetes.io/projected/0e6213f8-9948-47e9-88ac-7d7bdaddac26-kube-api-access-q6jq2\") pod \"dnsmasq-dns-759c6cc4df-s47tx\" (UID: \"0e6213f8-9948-47e9-88ac-7d7bdaddac26\") " pod="openstack/dnsmasq-dns-759c6cc4df-s47tx" Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.252162 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e6213f8-9948-47e9-88ac-7d7bdaddac26-config\") pod \"dnsmasq-dns-759c6cc4df-s47tx\" (UID: \"0e6213f8-9948-47e9-88ac-7d7bdaddac26\") " pod="openstack/dnsmasq-dns-759c6cc4df-s47tx" Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.252197 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e6213f8-9948-47e9-88ac-7d7bdaddac26-dns-svc\") pod \"dnsmasq-dns-759c6cc4df-s47tx\" (UID: \"0e6213f8-9948-47e9-88ac-7d7bdaddac26\") " pod="openstack/dnsmasq-dns-759c6cc4df-s47tx" Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.353822 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6jq2\" (UniqueName: \"kubernetes.io/projected/0e6213f8-9948-47e9-88ac-7d7bdaddac26-kube-api-access-q6jq2\") pod \"dnsmasq-dns-759c6cc4df-s47tx\" (UID: \"0e6213f8-9948-47e9-88ac-7d7bdaddac26\") " pod="openstack/dnsmasq-dns-759c6cc4df-s47tx" Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.354104 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e6213f8-9948-47e9-88ac-7d7bdaddac26-config\") pod \"dnsmasq-dns-759c6cc4df-s47tx\" (UID: \"0e6213f8-9948-47e9-88ac-7d7bdaddac26\") " pod="openstack/dnsmasq-dns-759c6cc4df-s47tx" Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.354166 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e6213f8-9948-47e9-88ac-7d7bdaddac26-dns-svc\") pod \"dnsmasq-dns-759c6cc4df-s47tx\" (UID: \"0e6213f8-9948-47e9-88ac-7d7bdaddac26\") " pod="openstack/dnsmasq-dns-759c6cc4df-s47tx" Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.355615 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e6213f8-9948-47e9-88ac-7d7bdaddac26-dns-svc\") pod \"dnsmasq-dns-759c6cc4df-s47tx\" (UID: \"0e6213f8-9948-47e9-88ac-7d7bdaddac26\") " pod="openstack/dnsmasq-dns-759c6cc4df-s47tx" Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.357113 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e6213f8-9948-47e9-88ac-7d7bdaddac26-config\") pod \"dnsmasq-dns-759c6cc4df-s47tx\" (UID: \"0e6213f8-9948-47e9-88ac-7d7bdaddac26\") " pod="openstack/dnsmasq-dns-759c6cc4df-s47tx" Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.378764 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6jq2\" (UniqueName: \"kubernetes.io/projected/0e6213f8-9948-47e9-88ac-7d7bdaddac26-kube-api-access-q6jq2\") pod \"dnsmasq-dns-759c6cc4df-s47tx\" (UID: \"0e6213f8-9948-47e9-88ac-7d7bdaddac26\") " pod="openstack/dnsmasq-dns-759c6cc4df-s47tx" Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.459267 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-759c6cc4df-s47tx" Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.635673 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f47fdfb89-xjlhn"] Nov 23 06:57:43 crc kubenswrapper[4559]: W1123 06:57:43.650827 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podebba232a_99bb_4f45_a0ce_3628f42dcc77.slice/crio-10bed732f0acea79aff1443fa19dd7bdf6eeede05443ef73df76cd16a1dde70f WatchSource:0}: Error finding container 10bed732f0acea79aff1443fa19dd7bdf6eeede05443ef73df76cd16a1dde70f: Status 404 returned error can't find the container with id 10bed732f0acea79aff1443fa19dd7bdf6eeede05443ef73df76cd16a1dde70f Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.840511 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-759c6cc4df-s47tx"] Nov 23 06:57:43 crc kubenswrapper[4559]: W1123 06:57:43.842510 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e6213f8_9948_47e9_88ac_7d7bdaddac26.slice/crio-117c6950fa5e853b90db417edea073284ebf5b1fa2b4d2bd374fb78495526e8f WatchSource:0}: Error finding container 117c6950fa5e853b90db417edea073284ebf5b1fa2b4d2bd374fb78495526e8f: Status 404 returned error can't find the container with id 117c6950fa5e853b90db417edea073284ebf5b1fa2b4d2bd374fb78495526e8f Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.970838 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.972095 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.975174 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.975414 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.975978 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.976242 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-lgflk" Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.976399 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.976532 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.976720 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 23 06:57:43 crc kubenswrapper[4559]: I1123 06:57:43.984427 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.065727 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.065802 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.065825 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.065948 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.065992 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.066008 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.066055 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd2cg\" (UniqueName: \"kubernetes.io/projected/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-kube-api-access-zd2cg\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.066201 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.066271 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.066299 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-config-data\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.066345 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.167389 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.167481 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.167513 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-config-data\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.167578 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.167602 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.167686 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.167708 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.167882 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.167920 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.167937 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.167945 4559 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.167967 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd2cg\" (UniqueName: \"kubernetes.io/projected/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-kube-api-access-zd2cg\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.169086 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.169121 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.169378 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-config-data\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.170036 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.172404 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.173808 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.173808 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.174149 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.175890 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.183973 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd2cg\" (UniqueName: \"kubernetes.io/projected/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-kube-api-access-zd2cg\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.185201 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.231325 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.239719 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.241843 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.245612 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.246262 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.246432 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.246544 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.246758 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-frpw9" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.246880 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.248509 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.268773 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/61e2f87a-db68-432a-8d13-94087d1afb17-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.269111 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/61e2f87a-db68-432a-8d13-94087d1afb17-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.269137 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9qg9\" (UniqueName: \"kubernetes.io/projected/61e2f87a-db68-432a-8d13-94087d1afb17-kube-api-access-s9qg9\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.269155 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/61e2f87a-db68-432a-8d13-94087d1afb17-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.269193 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/61e2f87a-db68-432a-8d13-94087d1afb17-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.269252 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/61e2f87a-db68-432a-8d13-94087d1afb17-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.269285 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.269313 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/61e2f87a-db68-432a-8d13-94087d1afb17-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.269536 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/61e2f87a-db68-432a-8d13-94087d1afb17-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.269597 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/61e2f87a-db68-432a-8d13-94087d1afb17-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.269663 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/61e2f87a-db68-432a-8d13-94087d1afb17-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.298251 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.370823 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/61e2f87a-db68-432a-8d13-94087d1afb17-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.371449 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/61e2f87a-db68-432a-8d13-94087d1afb17-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.371515 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/61e2f87a-db68-432a-8d13-94087d1afb17-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.371593 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9qg9\" (UniqueName: \"kubernetes.io/projected/61e2f87a-db68-432a-8d13-94087d1afb17-kube-api-access-s9qg9\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.371923 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/61e2f87a-db68-432a-8d13-94087d1afb17-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.372391 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/61e2f87a-db68-432a-8d13-94087d1afb17-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.372430 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/61e2f87a-db68-432a-8d13-94087d1afb17-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.372469 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.372559 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/61e2f87a-db68-432a-8d13-94087d1afb17-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.372588 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/61e2f87a-db68-432a-8d13-94087d1afb17-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.372636 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/61e2f87a-db68-432a-8d13-94087d1afb17-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.372693 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/61e2f87a-db68-432a-8d13-94087d1afb17-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.374253 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/61e2f87a-db68-432a-8d13-94087d1afb17-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.374591 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/61e2f87a-db68-432a-8d13-94087d1afb17-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.374954 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/61e2f87a-db68-432a-8d13-94087d1afb17-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.375055 4559 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.375812 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/61e2f87a-db68-432a-8d13-94087d1afb17-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.376048 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/61e2f87a-db68-432a-8d13-94087d1afb17-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.382241 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/61e2f87a-db68-432a-8d13-94087d1afb17-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.383098 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/61e2f87a-db68-432a-8d13-94087d1afb17-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.391133 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9qg9\" (UniqueName: \"kubernetes.io/projected/61e2f87a-db68-432a-8d13-94087d1afb17-kube-api-access-s9qg9\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.391335 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/61e2f87a-db68-432a-8d13-94087d1afb17-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.400163 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.563413 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.622977 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f47fdfb89-xjlhn" event={"ID":"ebba232a-99bb-4f45-a0ce-3628f42dcc77","Type":"ContainerStarted","Data":"10bed732f0acea79aff1443fa19dd7bdf6eeede05443ef73df76cd16a1dde70f"} Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.624553 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-759c6cc4df-s47tx" event={"ID":"0e6213f8-9948-47e9-88ac-7d7bdaddac26","Type":"ContainerStarted","Data":"117c6950fa5e853b90db417edea073284ebf5b1fa2b4d2bd374fb78495526e8f"} Nov 23 06:57:44 crc kubenswrapper[4559]: I1123 06:57:44.685780 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.014228 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 23 06:57:45 crc kubenswrapper[4559]: W1123 06:57:45.021785 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61e2f87a_db68_432a_8d13_94087d1afb17.slice/crio-370d0ad9a0809d04c728aaedbc819ffca79441a16c588bc68e81736ed67ad218 WatchSource:0}: Error finding container 370d0ad9a0809d04c728aaedbc819ffca79441a16c588bc68e81736ed67ad218: Status 404 returned error can't find the container with id 370d0ad9a0809d04c728aaedbc819ffca79441a16c588bc68e81736ed67ad218 Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.101695 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kb4bw"] Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.107185 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kb4bw" Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.150035 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kb4bw"] Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.195420 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99psz\" (UniqueName: \"kubernetes.io/projected/d6f08db9-323e-43dd-b925-1eea5dd88345-kube-api-access-99psz\") pod \"certified-operators-kb4bw\" (UID: \"d6f08db9-323e-43dd-b925-1eea5dd88345\") " pod="openshift-marketplace/certified-operators-kb4bw" Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.195526 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6f08db9-323e-43dd-b925-1eea5dd88345-utilities\") pod \"certified-operators-kb4bw\" (UID: \"d6f08db9-323e-43dd-b925-1eea5dd88345\") " pod="openshift-marketplace/certified-operators-kb4bw" Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.195654 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6f08db9-323e-43dd-b925-1eea5dd88345-catalog-content\") pod \"certified-operators-kb4bw\" (UID: \"d6f08db9-323e-43dd-b925-1eea5dd88345\") " pod="openshift-marketplace/certified-operators-kb4bw" Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.297207 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6f08db9-323e-43dd-b925-1eea5dd88345-utilities\") pod \"certified-operators-kb4bw\" (UID: \"d6f08db9-323e-43dd-b925-1eea5dd88345\") " pod="openshift-marketplace/certified-operators-kb4bw" Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.297344 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6f08db9-323e-43dd-b925-1eea5dd88345-catalog-content\") pod \"certified-operators-kb4bw\" (UID: \"d6f08db9-323e-43dd-b925-1eea5dd88345\") " pod="openshift-marketplace/certified-operators-kb4bw" Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.297409 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99psz\" (UniqueName: \"kubernetes.io/projected/d6f08db9-323e-43dd-b925-1eea5dd88345-kube-api-access-99psz\") pod \"certified-operators-kb4bw\" (UID: \"d6f08db9-323e-43dd-b925-1eea5dd88345\") " pod="openshift-marketplace/certified-operators-kb4bw" Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.297856 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6f08db9-323e-43dd-b925-1eea5dd88345-utilities\") pod \"certified-operators-kb4bw\" (UID: \"d6f08db9-323e-43dd-b925-1eea5dd88345\") " pod="openshift-marketplace/certified-operators-kb4bw" Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.298077 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6f08db9-323e-43dd-b925-1eea5dd88345-catalog-content\") pod \"certified-operators-kb4bw\" (UID: \"d6f08db9-323e-43dd-b925-1eea5dd88345\") " pod="openshift-marketplace/certified-operators-kb4bw" Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.318500 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99psz\" (UniqueName: \"kubernetes.io/projected/d6f08db9-323e-43dd-b925-1eea5dd88345-kube-api-access-99psz\") pod \"certified-operators-kb4bw\" (UID: \"d6f08db9-323e-43dd-b925-1eea5dd88345\") " pod="openshift-marketplace/certified-operators-kb4bw" Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.434817 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kb4bw" Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.632041 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"61e2f87a-db68-432a-8d13-94087d1afb17","Type":"ContainerStarted","Data":"370d0ad9a0809d04c728aaedbc819ffca79441a16c588bc68e81736ed67ad218"} Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.648245 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20","Type":"ContainerStarted","Data":"7458fff170fb3e48c0b29bbb98727c892d3f89f3718b88553d0f95c8c0752c96"} Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.797453 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.798702 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.804310 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.805515 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.805616 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.805733 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-nr597" Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.809798 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.810237 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.913299 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e097ab60-4653-4038-b6de-40b55936565c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"e097ab60-4653-4038-b6de-40b55936565c\") " pod="openstack/openstack-galera-0" Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.913572 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e097ab60-4653-4038-b6de-40b55936565c-config-data-default\") pod \"openstack-galera-0\" (UID: \"e097ab60-4653-4038-b6de-40b55936565c\") " pod="openstack/openstack-galera-0" Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.913592 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e097ab60-4653-4038-b6de-40b55936565c-kolla-config\") pod \"openstack-galera-0\" (UID: \"e097ab60-4653-4038-b6de-40b55936565c\") " pod="openstack/openstack-galera-0" Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.913727 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e097ab60-4653-4038-b6de-40b55936565c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"e097ab60-4653-4038-b6de-40b55936565c\") " pod="openstack/openstack-galera-0" Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.913746 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e097ab60-4653-4038-b6de-40b55936565c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"e097ab60-4653-4038-b6de-40b55936565c\") " pod="openstack/openstack-galera-0" Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.913766 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"e097ab60-4653-4038-b6de-40b55936565c\") " pod="openstack/openstack-galera-0" Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.913819 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w7q9\" (UniqueName: \"kubernetes.io/projected/e097ab60-4653-4038-b6de-40b55936565c-kube-api-access-8w7q9\") pod \"openstack-galera-0\" (UID: \"e097ab60-4653-4038-b6de-40b55936565c\") " pod="openstack/openstack-galera-0" Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.913887 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e097ab60-4653-4038-b6de-40b55936565c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"e097ab60-4653-4038-b6de-40b55936565c\") " pod="openstack/openstack-galera-0" Nov 23 06:57:45 crc kubenswrapper[4559]: I1123 06:57:45.924709 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kb4bw"] Nov 23 06:57:45 crc kubenswrapper[4559]: W1123 06:57:45.947606 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6f08db9_323e_43dd_b925_1eea5dd88345.slice/crio-7a39a4b1f615be0f2a41503eef627e7e7b74d67967748ec20f08cb31635d27e4 WatchSource:0}: Error finding container 7a39a4b1f615be0f2a41503eef627e7e7b74d67967748ec20f08cb31635d27e4: Status 404 returned error can't find the container with id 7a39a4b1f615be0f2a41503eef627e7e7b74d67967748ec20f08cb31635d27e4 Nov 23 06:57:46 crc kubenswrapper[4559]: I1123 06:57:46.015258 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e097ab60-4653-4038-b6de-40b55936565c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"e097ab60-4653-4038-b6de-40b55936565c\") " pod="openstack/openstack-galera-0" Nov 23 06:57:46 crc kubenswrapper[4559]: I1123 06:57:46.015311 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e097ab60-4653-4038-b6de-40b55936565c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"e097ab60-4653-4038-b6de-40b55936565c\") " pod="openstack/openstack-galera-0" Nov 23 06:57:46 crc kubenswrapper[4559]: I1123 06:57:46.015332 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"e097ab60-4653-4038-b6de-40b55936565c\") " pod="openstack/openstack-galera-0" Nov 23 06:57:46 crc kubenswrapper[4559]: I1123 06:57:46.015419 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w7q9\" (UniqueName: \"kubernetes.io/projected/e097ab60-4653-4038-b6de-40b55936565c-kube-api-access-8w7q9\") pod \"openstack-galera-0\" (UID: \"e097ab60-4653-4038-b6de-40b55936565c\") " pod="openstack/openstack-galera-0" Nov 23 06:57:46 crc kubenswrapper[4559]: I1123 06:57:46.015484 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e097ab60-4653-4038-b6de-40b55936565c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"e097ab60-4653-4038-b6de-40b55936565c\") " pod="openstack/openstack-galera-0" Nov 23 06:57:46 crc kubenswrapper[4559]: I1123 06:57:46.015569 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e097ab60-4653-4038-b6de-40b55936565c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"e097ab60-4653-4038-b6de-40b55936565c\") " pod="openstack/openstack-galera-0" Nov 23 06:57:46 crc kubenswrapper[4559]: I1123 06:57:46.015603 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e097ab60-4653-4038-b6de-40b55936565c-config-data-default\") pod \"openstack-galera-0\" (UID: \"e097ab60-4653-4038-b6de-40b55936565c\") " pod="openstack/openstack-galera-0" Nov 23 06:57:46 crc kubenswrapper[4559]: I1123 06:57:46.015622 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e097ab60-4653-4038-b6de-40b55936565c-kolla-config\") pod \"openstack-galera-0\" (UID: \"e097ab60-4653-4038-b6de-40b55936565c\") " pod="openstack/openstack-galera-0" Nov 23 06:57:46 crc kubenswrapper[4559]: I1123 06:57:46.017707 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e097ab60-4653-4038-b6de-40b55936565c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"e097ab60-4653-4038-b6de-40b55936565c\") " pod="openstack/openstack-galera-0" Nov 23 06:57:46 crc kubenswrapper[4559]: I1123 06:57:46.017917 4559 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"e097ab60-4653-4038-b6de-40b55936565c\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/openstack-galera-0" Nov 23 06:57:46 crc kubenswrapper[4559]: I1123 06:57:46.018119 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e097ab60-4653-4038-b6de-40b55936565c-kolla-config\") pod \"openstack-galera-0\" (UID: \"e097ab60-4653-4038-b6de-40b55936565c\") " pod="openstack/openstack-galera-0" Nov 23 06:57:46 crc kubenswrapper[4559]: I1123 06:57:46.018737 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e097ab60-4653-4038-b6de-40b55936565c-config-data-default\") pod \"openstack-galera-0\" (UID: \"e097ab60-4653-4038-b6de-40b55936565c\") " pod="openstack/openstack-galera-0" Nov 23 06:57:46 crc kubenswrapper[4559]: I1123 06:57:46.019246 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e097ab60-4653-4038-b6de-40b55936565c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"e097ab60-4653-4038-b6de-40b55936565c\") " pod="openstack/openstack-galera-0" Nov 23 06:57:46 crc kubenswrapper[4559]: I1123 06:57:46.022913 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e097ab60-4653-4038-b6de-40b55936565c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"e097ab60-4653-4038-b6de-40b55936565c\") " pod="openstack/openstack-galera-0" Nov 23 06:57:46 crc kubenswrapper[4559]: I1123 06:57:46.023383 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e097ab60-4653-4038-b6de-40b55936565c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"e097ab60-4653-4038-b6de-40b55936565c\") " pod="openstack/openstack-galera-0" Nov 23 06:57:46 crc kubenswrapper[4559]: I1123 06:57:46.033367 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w7q9\" (UniqueName: \"kubernetes.io/projected/e097ab60-4653-4038-b6de-40b55936565c-kube-api-access-8w7q9\") pod \"openstack-galera-0\" (UID: \"e097ab60-4653-4038-b6de-40b55936565c\") " pod="openstack/openstack-galera-0" Nov 23 06:57:46 crc kubenswrapper[4559]: I1123 06:57:46.041989 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"e097ab60-4653-4038-b6de-40b55936565c\") " pod="openstack/openstack-galera-0" Nov 23 06:57:46 crc kubenswrapper[4559]: I1123 06:57:46.124192 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 23 06:57:46 crc kubenswrapper[4559]: I1123 06:57:46.535091 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 23 06:57:46 crc kubenswrapper[4559]: W1123 06:57:46.542002 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode097ab60_4653_4038_b6de_40b55936565c.slice/crio-998091a35af0babed274f256e2281954fa5bf5ae34c3f20fc602387f0111a9f4 WatchSource:0}: Error finding container 998091a35af0babed274f256e2281954fa5bf5ae34c3f20fc602387f0111a9f4: Status 404 returned error can't find the container with id 998091a35af0babed274f256e2281954fa5bf5ae34c3f20fc602387f0111a9f4 Nov 23 06:57:46 crc kubenswrapper[4559]: I1123 06:57:46.655884 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"e097ab60-4653-4038-b6de-40b55936565c","Type":"ContainerStarted","Data":"998091a35af0babed274f256e2281954fa5bf5ae34c3f20fc602387f0111a9f4"} Nov 23 06:57:46 crc kubenswrapper[4559]: I1123 06:57:46.657821 4559 generic.go:334] "Generic (PLEG): container finished" podID="d6f08db9-323e-43dd-b925-1eea5dd88345" containerID="c12d69435ac9c13967e865a1980550ffe3c0f8e3660ba9b21a9f513ff88b6c14" exitCode=0 Nov 23 06:57:46 crc kubenswrapper[4559]: I1123 06:57:46.657877 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kb4bw" event={"ID":"d6f08db9-323e-43dd-b925-1eea5dd88345","Type":"ContainerDied","Data":"c12d69435ac9c13967e865a1980550ffe3c0f8e3660ba9b21a9f513ff88b6c14"} Nov 23 06:57:46 crc kubenswrapper[4559]: I1123 06:57:46.657938 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kb4bw" event={"ID":"d6f08db9-323e-43dd-b925-1eea5dd88345","Type":"ContainerStarted","Data":"7a39a4b1f615be0f2a41503eef627e7e7b74d67967748ec20f08cb31635d27e4"} Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.113991 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.116925 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.122861 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.127546 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-x6kht" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.127694 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.128320 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.128814 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.244259 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01dff7d1-bf4c-4d88-900c-99de62949ac7-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"01dff7d1-bf4c-4d88-900c-99de62949ac7\") " pod="openstack/openstack-cell1-galera-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.244330 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/01dff7d1-bf4c-4d88-900c-99de62949ac7-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"01dff7d1-bf4c-4d88-900c-99de62949ac7\") " pod="openstack/openstack-cell1-galera-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.244361 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"01dff7d1-bf4c-4d88-900c-99de62949ac7\") " pod="openstack/openstack-cell1-galera-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.244472 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/01dff7d1-bf4c-4d88-900c-99de62949ac7-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"01dff7d1-bf4c-4d88-900c-99de62949ac7\") " pod="openstack/openstack-cell1-galera-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.244532 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79s44\" (UniqueName: \"kubernetes.io/projected/01dff7d1-bf4c-4d88-900c-99de62949ac7-kube-api-access-79s44\") pod \"openstack-cell1-galera-0\" (UID: \"01dff7d1-bf4c-4d88-900c-99de62949ac7\") " pod="openstack/openstack-cell1-galera-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.244553 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/01dff7d1-bf4c-4d88-900c-99de62949ac7-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"01dff7d1-bf4c-4d88-900c-99de62949ac7\") " pod="openstack/openstack-cell1-galera-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.244593 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/01dff7d1-bf4c-4d88-900c-99de62949ac7-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"01dff7d1-bf4c-4d88-900c-99de62949ac7\") " pod="openstack/openstack-cell1-galera-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.244610 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01dff7d1-bf4c-4d88-900c-99de62949ac7-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"01dff7d1-bf4c-4d88-900c-99de62949ac7\") " pod="openstack/openstack-cell1-galera-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.346371 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01dff7d1-bf4c-4d88-900c-99de62949ac7-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"01dff7d1-bf4c-4d88-900c-99de62949ac7\") " pod="openstack/openstack-cell1-galera-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.346483 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/01dff7d1-bf4c-4d88-900c-99de62949ac7-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"01dff7d1-bf4c-4d88-900c-99de62949ac7\") " pod="openstack/openstack-cell1-galera-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.346514 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"01dff7d1-bf4c-4d88-900c-99de62949ac7\") " pod="openstack/openstack-cell1-galera-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.346536 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/01dff7d1-bf4c-4d88-900c-99de62949ac7-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"01dff7d1-bf4c-4d88-900c-99de62949ac7\") " pod="openstack/openstack-cell1-galera-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.346653 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79s44\" (UniqueName: \"kubernetes.io/projected/01dff7d1-bf4c-4d88-900c-99de62949ac7-kube-api-access-79s44\") pod \"openstack-cell1-galera-0\" (UID: \"01dff7d1-bf4c-4d88-900c-99de62949ac7\") " pod="openstack/openstack-cell1-galera-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.346679 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/01dff7d1-bf4c-4d88-900c-99de62949ac7-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"01dff7d1-bf4c-4d88-900c-99de62949ac7\") " pod="openstack/openstack-cell1-galera-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.346762 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01dff7d1-bf4c-4d88-900c-99de62949ac7-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"01dff7d1-bf4c-4d88-900c-99de62949ac7\") " pod="openstack/openstack-cell1-galera-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.347002 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/01dff7d1-bf4c-4d88-900c-99de62949ac7-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"01dff7d1-bf4c-4d88-900c-99de62949ac7\") " pod="openstack/openstack-cell1-galera-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.353577 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/01dff7d1-bf4c-4d88-900c-99de62949ac7-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"01dff7d1-bf4c-4d88-900c-99de62949ac7\") " pod="openstack/openstack-cell1-galera-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.356138 4559 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"01dff7d1-bf4c-4d88-900c-99de62949ac7\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-cell1-galera-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.356920 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01dff7d1-bf4c-4d88-900c-99de62949ac7-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"01dff7d1-bf4c-4d88-900c-99de62949ac7\") " pod="openstack/openstack-cell1-galera-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.357781 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01dff7d1-bf4c-4d88-900c-99de62949ac7-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"01dff7d1-bf4c-4d88-900c-99de62949ac7\") " pod="openstack/openstack-cell1-galera-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.358077 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/01dff7d1-bf4c-4d88-900c-99de62949ac7-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"01dff7d1-bf4c-4d88-900c-99de62949ac7\") " pod="openstack/openstack-cell1-galera-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.362749 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/01dff7d1-bf4c-4d88-900c-99de62949ac7-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"01dff7d1-bf4c-4d88-900c-99de62949ac7\") " pod="openstack/openstack-cell1-galera-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.363317 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/01dff7d1-bf4c-4d88-900c-99de62949ac7-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"01dff7d1-bf4c-4d88-900c-99de62949ac7\") " pod="openstack/openstack-cell1-galera-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.370178 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79s44\" (UniqueName: \"kubernetes.io/projected/01dff7d1-bf4c-4d88-900c-99de62949ac7-kube-api-access-79s44\") pod \"openstack-cell1-galera-0\" (UID: \"01dff7d1-bf4c-4d88-900c-99de62949ac7\") " pod="openstack/openstack-cell1-galera-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.383156 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"01dff7d1-bf4c-4d88-900c-99de62949ac7\") " pod="openstack/openstack-cell1-galera-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.438106 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.662764 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.664601 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.669542 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-nfv2b" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.669711 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.669819 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.676380 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.699959 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kb4bw" event={"ID":"d6f08db9-323e-43dd-b925-1eea5dd88345","Type":"ContainerStarted","Data":"0a77d5d8f83d5f50554f687a003de50e674d73f27e63c9d6af603da4f236b674"} Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.761512 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/de0cb0c4-ffde-44c5-adb9-7aea0692c1b5-memcached-tls-certs\") pod \"memcached-0\" (UID: \"de0cb0c4-ffde-44c5-adb9-7aea0692c1b5\") " pod="openstack/memcached-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.761711 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/de0cb0c4-ffde-44c5-adb9-7aea0692c1b5-config-data\") pod \"memcached-0\" (UID: \"de0cb0c4-ffde-44c5-adb9-7aea0692c1b5\") " pod="openstack/memcached-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.761998 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de0cb0c4-ffde-44c5-adb9-7aea0692c1b5-combined-ca-bundle\") pod \"memcached-0\" (UID: \"de0cb0c4-ffde-44c5-adb9-7aea0692c1b5\") " pod="openstack/memcached-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.762053 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/de0cb0c4-ffde-44c5-adb9-7aea0692c1b5-kolla-config\") pod \"memcached-0\" (UID: \"de0cb0c4-ffde-44c5-adb9-7aea0692c1b5\") " pod="openstack/memcached-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.762070 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6bqj\" (UniqueName: \"kubernetes.io/projected/de0cb0c4-ffde-44c5-adb9-7aea0692c1b5-kube-api-access-b6bqj\") pod \"memcached-0\" (UID: \"de0cb0c4-ffde-44c5-adb9-7aea0692c1b5\") " pod="openstack/memcached-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.865271 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de0cb0c4-ffde-44c5-adb9-7aea0692c1b5-combined-ca-bundle\") pod \"memcached-0\" (UID: \"de0cb0c4-ffde-44c5-adb9-7aea0692c1b5\") " pod="openstack/memcached-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.865338 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/de0cb0c4-ffde-44c5-adb9-7aea0692c1b5-kolla-config\") pod \"memcached-0\" (UID: \"de0cb0c4-ffde-44c5-adb9-7aea0692c1b5\") " pod="openstack/memcached-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.865357 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6bqj\" (UniqueName: \"kubernetes.io/projected/de0cb0c4-ffde-44c5-adb9-7aea0692c1b5-kube-api-access-b6bqj\") pod \"memcached-0\" (UID: \"de0cb0c4-ffde-44c5-adb9-7aea0692c1b5\") " pod="openstack/memcached-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.865397 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/de0cb0c4-ffde-44c5-adb9-7aea0692c1b5-memcached-tls-certs\") pod \"memcached-0\" (UID: \"de0cb0c4-ffde-44c5-adb9-7aea0692c1b5\") " pod="openstack/memcached-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.865498 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/de0cb0c4-ffde-44c5-adb9-7aea0692c1b5-config-data\") pod \"memcached-0\" (UID: \"de0cb0c4-ffde-44c5-adb9-7aea0692c1b5\") " pod="openstack/memcached-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.867199 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/de0cb0c4-ffde-44c5-adb9-7aea0692c1b5-kolla-config\") pod \"memcached-0\" (UID: \"de0cb0c4-ffde-44c5-adb9-7aea0692c1b5\") " pod="openstack/memcached-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.870086 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/de0cb0c4-ffde-44c5-adb9-7aea0692c1b5-config-data\") pod \"memcached-0\" (UID: \"de0cb0c4-ffde-44c5-adb9-7aea0692c1b5\") " pod="openstack/memcached-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.873222 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/de0cb0c4-ffde-44c5-adb9-7aea0692c1b5-memcached-tls-certs\") pod \"memcached-0\" (UID: \"de0cb0c4-ffde-44c5-adb9-7aea0692c1b5\") " pod="openstack/memcached-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.877946 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de0cb0c4-ffde-44c5-adb9-7aea0692c1b5-combined-ca-bundle\") pod \"memcached-0\" (UID: \"de0cb0c4-ffde-44c5-adb9-7aea0692c1b5\") " pod="openstack/memcached-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.888066 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6bqj\" (UniqueName: \"kubernetes.io/projected/de0cb0c4-ffde-44c5-adb9-7aea0692c1b5-kube-api-access-b6bqj\") pod \"memcached-0\" (UID: \"de0cb0c4-ffde-44c5-adb9-7aea0692c1b5\") " pod="openstack/memcached-0" Nov 23 06:57:47 crc kubenswrapper[4559]: I1123 06:57:47.994495 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 23 06:57:48 crc kubenswrapper[4559]: I1123 06:57:48.197336 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 23 06:57:48 crc kubenswrapper[4559]: W1123 06:57:48.203977 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01dff7d1_bf4c_4d88_900c_99de62949ac7.slice/crio-1ff84d1e6566fb53a83d4ba1a5b16e6357906674d9182ef85bea75c70a24ddb3 WatchSource:0}: Error finding container 1ff84d1e6566fb53a83d4ba1a5b16e6357906674d9182ef85bea75c70a24ddb3: Status 404 returned error can't find the container with id 1ff84d1e6566fb53a83d4ba1a5b16e6357906674d9182ef85bea75c70a24ddb3 Nov 23 06:57:48 crc kubenswrapper[4559]: I1123 06:57:48.413550 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 23 06:57:48 crc kubenswrapper[4559]: I1123 06:57:48.714807 4559 generic.go:334] "Generic (PLEG): container finished" podID="d6f08db9-323e-43dd-b925-1eea5dd88345" containerID="0a77d5d8f83d5f50554f687a003de50e674d73f27e63c9d6af603da4f236b674" exitCode=0 Nov 23 06:57:48 crc kubenswrapper[4559]: I1123 06:57:48.714904 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kb4bw" event={"ID":"d6f08db9-323e-43dd-b925-1eea5dd88345","Type":"ContainerDied","Data":"0a77d5d8f83d5f50554f687a003de50e674d73f27e63c9d6af603da4f236b674"} Nov 23 06:57:48 crc kubenswrapper[4559]: I1123 06:57:48.716793 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"de0cb0c4-ffde-44c5-adb9-7aea0692c1b5","Type":"ContainerStarted","Data":"0e42b9637d3164d2f6b32ee23142f02921372e1076f5b67b3779471425243bd8"} Nov 23 06:57:48 crc kubenswrapper[4559]: I1123 06:57:48.719162 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"01dff7d1-bf4c-4d88-900c-99de62949ac7","Type":"ContainerStarted","Data":"1ff84d1e6566fb53a83d4ba1a5b16e6357906674d9182ef85bea75c70a24ddb3"} Nov 23 06:57:49 crc kubenswrapper[4559]: I1123 06:57:49.171418 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 06:57:49 crc kubenswrapper[4559]: I1123 06:57:49.172715 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 23 06:57:49 crc kubenswrapper[4559]: I1123 06:57:49.175471 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-rdzdv" Nov 23 06:57:49 crc kubenswrapper[4559]: I1123 06:57:49.176673 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 06:57:49 crc kubenswrapper[4559]: I1123 06:57:49.310633 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nrt6\" (UniqueName: \"kubernetes.io/projected/f306bc71-8c65-4236-b439-e3af9c495a15-kube-api-access-4nrt6\") pod \"kube-state-metrics-0\" (UID: \"f306bc71-8c65-4236-b439-e3af9c495a15\") " pod="openstack/kube-state-metrics-0" Nov 23 06:57:49 crc kubenswrapper[4559]: I1123 06:57:49.412786 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nrt6\" (UniqueName: \"kubernetes.io/projected/f306bc71-8c65-4236-b439-e3af9c495a15-kube-api-access-4nrt6\") pod \"kube-state-metrics-0\" (UID: \"f306bc71-8c65-4236-b439-e3af9c495a15\") " pod="openstack/kube-state-metrics-0" Nov 23 06:57:49 crc kubenswrapper[4559]: I1123 06:57:49.435600 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nrt6\" (UniqueName: \"kubernetes.io/projected/f306bc71-8c65-4236-b439-e3af9c495a15-kube-api-access-4nrt6\") pod \"kube-state-metrics-0\" (UID: \"f306bc71-8c65-4236-b439-e3af9c495a15\") " pod="openstack/kube-state-metrics-0" Nov 23 06:57:49 crc kubenswrapper[4559]: I1123 06:57:49.493077 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 23 06:57:49 crc kubenswrapper[4559]: I1123 06:57:49.736696 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kb4bw" event={"ID":"d6f08db9-323e-43dd-b925-1eea5dd88345","Type":"ContainerStarted","Data":"39e183bd477c83b9156f9160c9674fa9fb44501acbb4b1d1ae22fe79ecce18d1"} Nov 23 06:57:49 crc kubenswrapper[4559]: I1123 06:57:49.754740 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kb4bw" podStartSLOduration=2.19414219 podStartE2EDuration="4.754722645s" podCreationTimestamp="2025-11-23 06:57:45 +0000 UTC" firstStartedPulling="2025-11-23 06:57:46.65962394 +0000 UTC m=+768.681609554" lastFinishedPulling="2025-11-23 06:57:49.220204405 +0000 UTC m=+771.242190009" observedRunningTime="2025-11-23 06:57:49.751315268 +0000 UTC m=+771.773300881" watchObservedRunningTime="2025-11-23 06:57:49.754722645 +0000 UTC m=+771.776708259" Nov 23 06:57:49 crc kubenswrapper[4559]: I1123 06:57:49.953133 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 06:57:49 crc kubenswrapper[4559]: W1123 06:57:49.966531 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf306bc71_8c65_4236_b439_e3af9c495a15.slice/crio-a82e6e70536b9910866670d1d9955a3e45d2d9781f262ee685560c24e30f92ce WatchSource:0}: Error finding container a82e6e70536b9910866670d1d9955a3e45d2d9781f262ee685560c24e30f92ce: Status 404 returned error can't find the container with id a82e6e70536b9910866670d1d9955a3e45d2d9781f262ee685560c24e30f92ce Nov 23 06:57:50 crc kubenswrapper[4559]: I1123 06:57:50.758050 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f306bc71-8c65-4236-b439-e3af9c495a15","Type":"ContainerStarted","Data":"a82e6e70536b9910866670d1d9955a3e45d2d9781f262ee685560c24e30f92ce"} Nov 23 06:57:52 crc kubenswrapper[4559]: I1123 06:57:52.777576 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f306bc71-8c65-4236-b439-e3af9c495a15","Type":"ContainerStarted","Data":"c75a09b9967ab3f8d926027629a6b224d1d1615027f5f57d8d16bb2c450be721"} Nov 23 06:57:52 crc kubenswrapper[4559]: I1123 06:57:52.783810 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 23 06:57:52 crc kubenswrapper[4559]: I1123 06:57:52.799225 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.8599611390000002 podStartE2EDuration="3.799206903s" podCreationTimestamp="2025-11-23 06:57:49 +0000 UTC" firstStartedPulling="2025-11-23 06:57:49.970988663 +0000 UTC m=+771.992974276" lastFinishedPulling="2025-11-23 06:57:51.910234427 +0000 UTC m=+773.932220040" observedRunningTime="2025-11-23 06:57:52.798519522 +0000 UTC m=+774.820505135" watchObservedRunningTime="2025-11-23 06:57:52.799206903 +0000 UTC m=+774.821192518" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.075186 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-blj85"] Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.076556 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-blj85" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.078219 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-7c7dj" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.078673 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.078819 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.105722 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-jg7hk"] Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.112406 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-jg7hk" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.116193 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-blj85"] Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.122911 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-jg7hk"] Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.179564 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/20662e83-ba79-4c97-80fb-98fda28c1149-var-log-ovn\") pod \"ovn-controller-blj85\" (UID: \"20662e83-ba79-4c97-80fb-98fda28c1149\") " pod="openstack/ovn-controller-blj85" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.179738 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20662e83-ba79-4c97-80fb-98fda28c1149-combined-ca-bundle\") pod \"ovn-controller-blj85\" (UID: \"20662e83-ba79-4c97-80fb-98fda28c1149\") " pod="openstack/ovn-controller-blj85" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.179788 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20662e83-ba79-4c97-80fb-98fda28c1149-scripts\") pod \"ovn-controller-blj85\" (UID: \"20662e83-ba79-4c97-80fb-98fda28c1149\") " pod="openstack/ovn-controller-blj85" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.179853 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmsq8\" (UniqueName: \"kubernetes.io/projected/20662e83-ba79-4c97-80fb-98fda28c1149-kube-api-access-jmsq8\") pod \"ovn-controller-blj85\" (UID: \"20662e83-ba79-4c97-80fb-98fda28c1149\") " pod="openstack/ovn-controller-blj85" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.179905 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/20662e83-ba79-4c97-80fb-98fda28c1149-var-run\") pod \"ovn-controller-blj85\" (UID: \"20662e83-ba79-4c97-80fb-98fda28c1149\") " pod="openstack/ovn-controller-blj85" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.179976 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/20662e83-ba79-4c97-80fb-98fda28c1149-var-run-ovn\") pod \"ovn-controller-blj85\" (UID: \"20662e83-ba79-4c97-80fb-98fda28c1149\") " pod="openstack/ovn-controller-blj85" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.180030 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/20662e83-ba79-4c97-80fb-98fda28c1149-ovn-controller-tls-certs\") pod \"ovn-controller-blj85\" (UID: \"20662e83-ba79-4c97-80fb-98fda28c1149\") " pod="openstack/ovn-controller-blj85" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.281387 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d534f227-972b-4a85-aeb5-8f1d226352e1-var-log\") pod \"ovn-controller-ovs-jg7hk\" (UID: \"d534f227-972b-4a85-aeb5-8f1d226352e1\") " pod="openstack/ovn-controller-ovs-jg7hk" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.281464 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/20662e83-ba79-4c97-80fb-98fda28c1149-var-log-ovn\") pod \"ovn-controller-blj85\" (UID: \"20662e83-ba79-4c97-80fb-98fda28c1149\") " pod="openstack/ovn-controller-blj85" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.281488 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d534f227-972b-4a85-aeb5-8f1d226352e1-var-run\") pod \"ovn-controller-ovs-jg7hk\" (UID: \"d534f227-972b-4a85-aeb5-8f1d226352e1\") " pod="openstack/ovn-controller-ovs-jg7hk" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.281557 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7lc2\" (UniqueName: \"kubernetes.io/projected/d534f227-972b-4a85-aeb5-8f1d226352e1-kube-api-access-t7lc2\") pod \"ovn-controller-ovs-jg7hk\" (UID: \"d534f227-972b-4a85-aeb5-8f1d226352e1\") " pod="openstack/ovn-controller-ovs-jg7hk" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.281600 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20662e83-ba79-4c97-80fb-98fda28c1149-combined-ca-bundle\") pod \"ovn-controller-blj85\" (UID: \"20662e83-ba79-4c97-80fb-98fda28c1149\") " pod="openstack/ovn-controller-blj85" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.281692 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d534f227-972b-4a85-aeb5-8f1d226352e1-scripts\") pod \"ovn-controller-ovs-jg7hk\" (UID: \"d534f227-972b-4a85-aeb5-8f1d226352e1\") " pod="openstack/ovn-controller-ovs-jg7hk" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.281729 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20662e83-ba79-4c97-80fb-98fda28c1149-scripts\") pod \"ovn-controller-blj85\" (UID: \"20662e83-ba79-4c97-80fb-98fda28c1149\") " pod="openstack/ovn-controller-blj85" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.281797 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d534f227-972b-4a85-aeb5-8f1d226352e1-etc-ovs\") pod \"ovn-controller-ovs-jg7hk\" (UID: \"d534f227-972b-4a85-aeb5-8f1d226352e1\") " pod="openstack/ovn-controller-ovs-jg7hk" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.281856 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d534f227-972b-4a85-aeb5-8f1d226352e1-var-lib\") pod \"ovn-controller-ovs-jg7hk\" (UID: \"d534f227-972b-4a85-aeb5-8f1d226352e1\") " pod="openstack/ovn-controller-ovs-jg7hk" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.281895 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmsq8\" (UniqueName: \"kubernetes.io/projected/20662e83-ba79-4c97-80fb-98fda28c1149-kube-api-access-jmsq8\") pod \"ovn-controller-blj85\" (UID: \"20662e83-ba79-4c97-80fb-98fda28c1149\") " pod="openstack/ovn-controller-blj85" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.281958 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/20662e83-ba79-4c97-80fb-98fda28c1149-var-run\") pod \"ovn-controller-blj85\" (UID: \"20662e83-ba79-4c97-80fb-98fda28c1149\") " pod="openstack/ovn-controller-blj85" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.282002 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/20662e83-ba79-4c97-80fb-98fda28c1149-var-run-ovn\") pod \"ovn-controller-blj85\" (UID: \"20662e83-ba79-4c97-80fb-98fda28c1149\") " pod="openstack/ovn-controller-blj85" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.282029 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/20662e83-ba79-4c97-80fb-98fda28c1149-ovn-controller-tls-certs\") pod \"ovn-controller-blj85\" (UID: \"20662e83-ba79-4c97-80fb-98fda28c1149\") " pod="openstack/ovn-controller-blj85" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.284511 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20662e83-ba79-4c97-80fb-98fda28c1149-scripts\") pod \"ovn-controller-blj85\" (UID: \"20662e83-ba79-4c97-80fb-98fda28c1149\") " pod="openstack/ovn-controller-blj85" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.284914 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/20662e83-ba79-4c97-80fb-98fda28c1149-var-run-ovn\") pod \"ovn-controller-blj85\" (UID: \"20662e83-ba79-4c97-80fb-98fda28c1149\") " pod="openstack/ovn-controller-blj85" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.285264 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/20662e83-ba79-4c97-80fb-98fda28c1149-var-log-ovn\") pod \"ovn-controller-blj85\" (UID: \"20662e83-ba79-4c97-80fb-98fda28c1149\") " pod="openstack/ovn-controller-blj85" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.286326 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/20662e83-ba79-4c97-80fb-98fda28c1149-var-run\") pod \"ovn-controller-blj85\" (UID: \"20662e83-ba79-4c97-80fb-98fda28c1149\") " pod="openstack/ovn-controller-blj85" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.288432 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/20662e83-ba79-4c97-80fb-98fda28c1149-ovn-controller-tls-certs\") pod \"ovn-controller-blj85\" (UID: \"20662e83-ba79-4c97-80fb-98fda28c1149\") " pod="openstack/ovn-controller-blj85" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.298862 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20662e83-ba79-4c97-80fb-98fda28c1149-combined-ca-bundle\") pod \"ovn-controller-blj85\" (UID: \"20662e83-ba79-4c97-80fb-98fda28c1149\") " pod="openstack/ovn-controller-blj85" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.302051 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmsq8\" (UniqueName: \"kubernetes.io/projected/20662e83-ba79-4c97-80fb-98fda28c1149-kube-api-access-jmsq8\") pod \"ovn-controller-blj85\" (UID: \"20662e83-ba79-4c97-80fb-98fda28c1149\") " pod="openstack/ovn-controller-blj85" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.383918 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d534f227-972b-4a85-aeb5-8f1d226352e1-var-log\") pod \"ovn-controller-ovs-jg7hk\" (UID: \"d534f227-972b-4a85-aeb5-8f1d226352e1\") " pod="openstack/ovn-controller-ovs-jg7hk" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.384004 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d534f227-972b-4a85-aeb5-8f1d226352e1-var-run\") pod \"ovn-controller-ovs-jg7hk\" (UID: \"d534f227-972b-4a85-aeb5-8f1d226352e1\") " pod="openstack/ovn-controller-ovs-jg7hk" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.384041 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7lc2\" (UniqueName: \"kubernetes.io/projected/d534f227-972b-4a85-aeb5-8f1d226352e1-kube-api-access-t7lc2\") pod \"ovn-controller-ovs-jg7hk\" (UID: \"d534f227-972b-4a85-aeb5-8f1d226352e1\") " pod="openstack/ovn-controller-ovs-jg7hk" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.384091 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d534f227-972b-4a85-aeb5-8f1d226352e1-scripts\") pod \"ovn-controller-ovs-jg7hk\" (UID: \"d534f227-972b-4a85-aeb5-8f1d226352e1\") " pod="openstack/ovn-controller-ovs-jg7hk" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.384116 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d534f227-972b-4a85-aeb5-8f1d226352e1-etc-ovs\") pod \"ovn-controller-ovs-jg7hk\" (UID: \"d534f227-972b-4a85-aeb5-8f1d226352e1\") " pod="openstack/ovn-controller-ovs-jg7hk" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.384146 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d534f227-972b-4a85-aeb5-8f1d226352e1-var-lib\") pod \"ovn-controller-ovs-jg7hk\" (UID: \"d534f227-972b-4a85-aeb5-8f1d226352e1\") " pod="openstack/ovn-controller-ovs-jg7hk" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.384516 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d534f227-972b-4a85-aeb5-8f1d226352e1-var-lib\") pod \"ovn-controller-ovs-jg7hk\" (UID: \"d534f227-972b-4a85-aeb5-8f1d226352e1\") " pod="openstack/ovn-controller-ovs-jg7hk" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.385972 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d534f227-972b-4a85-aeb5-8f1d226352e1-var-log\") pod \"ovn-controller-ovs-jg7hk\" (UID: \"d534f227-972b-4a85-aeb5-8f1d226352e1\") " pod="openstack/ovn-controller-ovs-jg7hk" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.386049 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d534f227-972b-4a85-aeb5-8f1d226352e1-var-run\") pod \"ovn-controller-ovs-jg7hk\" (UID: \"d534f227-972b-4a85-aeb5-8f1d226352e1\") " pod="openstack/ovn-controller-ovs-jg7hk" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.388371 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d534f227-972b-4a85-aeb5-8f1d226352e1-etc-ovs\") pod \"ovn-controller-ovs-jg7hk\" (UID: \"d534f227-972b-4a85-aeb5-8f1d226352e1\") " pod="openstack/ovn-controller-ovs-jg7hk" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.389069 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d534f227-972b-4a85-aeb5-8f1d226352e1-scripts\") pod \"ovn-controller-ovs-jg7hk\" (UID: \"d534f227-972b-4a85-aeb5-8f1d226352e1\") " pod="openstack/ovn-controller-ovs-jg7hk" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.400356 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7lc2\" (UniqueName: \"kubernetes.io/projected/d534f227-972b-4a85-aeb5-8f1d226352e1-kube-api-access-t7lc2\") pod \"ovn-controller-ovs-jg7hk\" (UID: \"d534f227-972b-4a85-aeb5-8f1d226352e1\") " pod="openstack/ovn-controller-ovs-jg7hk" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.412924 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-blj85" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.424532 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-jg7hk" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.969734 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.971141 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.973325 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.974252 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.974534 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.975133 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-jvw59" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.976808 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 23 06:57:53 crc kubenswrapper[4559]: I1123 06:57:53.983297 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 23 06:57:54 crc kubenswrapper[4559]: I1123 06:57:54.094787 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d19fe8b-239a-476a-9213-1eccbd58958e-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5d19fe8b-239a-476a-9213-1eccbd58958e\") " pod="openstack/ovsdbserver-nb-0" Nov 23 06:57:54 crc kubenswrapper[4559]: I1123 06:57:54.094854 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d19fe8b-239a-476a-9213-1eccbd58958e-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5d19fe8b-239a-476a-9213-1eccbd58958e\") " pod="openstack/ovsdbserver-nb-0" Nov 23 06:57:54 crc kubenswrapper[4559]: I1123 06:57:54.094977 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d19fe8b-239a-476a-9213-1eccbd58958e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"5d19fe8b-239a-476a-9213-1eccbd58958e\") " pod="openstack/ovsdbserver-nb-0" Nov 23 06:57:54 crc kubenswrapper[4559]: I1123 06:57:54.095016 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5d19fe8b-239a-476a-9213-1eccbd58958e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"5d19fe8b-239a-476a-9213-1eccbd58958e\") " pod="openstack/ovsdbserver-nb-0" Nov 23 06:57:54 crc kubenswrapper[4559]: I1123 06:57:54.095040 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7qmd\" (UniqueName: \"kubernetes.io/projected/5d19fe8b-239a-476a-9213-1eccbd58958e-kube-api-access-g7qmd\") pod \"ovsdbserver-nb-0\" (UID: \"5d19fe8b-239a-476a-9213-1eccbd58958e\") " pod="openstack/ovsdbserver-nb-0" Nov 23 06:57:54 crc kubenswrapper[4559]: I1123 06:57:54.095096 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d19fe8b-239a-476a-9213-1eccbd58958e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"5d19fe8b-239a-476a-9213-1eccbd58958e\") " pod="openstack/ovsdbserver-nb-0" Nov 23 06:57:54 crc kubenswrapper[4559]: I1123 06:57:54.095127 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d19fe8b-239a-476a-9213-1eccbd58958e-config\") pod \"ovsdbserver-nb-0\" (UID: \"5d19fe8b-239a-476a-9213-1eccbd58958e\") " pod="openstack/ovsdbserver-nb-0" Nov 23 06:57:54 crc kubenswrapper[4559]: I1123 06:57:54.095356 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"5d19fe8b-239a-476a-9213-1eccbd58958e\") " pod="openstack/ovsdbserver-nb-0" Nov 23 06:57:54 crc kubenswrapper[4559]: I1123 06:57:54.197204 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"5d19fe8b-239a-476a-9213-1eccbd58958e\") " pod="openstack/ovsdbserver-nb-0" Nov 23 06:57:54 crc kubenswrapper[4559]: I1123 06:57:54.197264 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d19fe8b-239a-476a-9213-1eccbd58958e-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5d19fe8b-239a-476a-9213-1eccbd58958e\") " pod="openstack/ovsdbserver-nb-0" Nov 23 06:57:54 crc kubenswrapper[4559]: I1123 06:57:54.197308 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d19fe8b-239a-476a-9213-1eccbd58958e-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5d19fe8b-239a-476a-9213-1eccbd58958e\") " pod="openstack/ovsdbserver-nb-0" Nov 23 06:57:54 crc kubenswrapper[4559]: I1123 06:57:54.197337 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d19fe8b-239a-476a-9213-1eccbd58958e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"5d19fe8b-239a-476a-9213-1eccbd58958e\") " pod="openstack/ovsdbserver-nb-0" Nov 23 06:57:54 crc kubenswrapper[4559]: I1123 06:57:54.197357 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5d19fe8b-239a-476a-9213-1eccbd58958e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"5d19fe8b-239a-476a-9213-1eccbd58958e\") " pod="openstack/ovsdbserver-nb-0" Nov 23 06:57:54 crc kubenswrapper[4559]: I1123 06:57:54.197376 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7qmd\" (UniqueName: \"kubernetes.io/projected/5d19fe8b-239a-476a-9213-1eccbd58958e-kube-api-access-g7qmd\") pod \"ovsdbserver-nb-0\" (UID: \"5d19fe8b-239a-476a-9213-1eccbd58958e\") " pod="openstack/ovsdbserver-nb-0" Nov 23 06:57:54 crc kubenswrapper[4559]: I1123 06:57:54.197410 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d19fe8b-239a-476a-9213-1eccbd58958e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"5d19fe8b-239a-476a-9213-1eccbd58958e\") " pod="openstack/ovsdbserver-nb-0" Nov 23 06:57:54 crc kubenswrapper[4559]: I1123 06:57:54.197434 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d19fe8b-239a-476a-9213-1eccbd58958e-config\") pod \"ovsdbserver-nb-0\" (UID: \"5d19fe8b-239a-476a-9213-1eccbd58958e\") " pod="openstack/ovsdbserver-nb-0" Nov 23 06:57:54 crc kubenswrapper[4559]: I1123 06:57:54.198242 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d19fe8b-239a-476a-9213-1eccbd58958e-config\") pod \"ovsdbserver-nb-0\" (UID: \"5d19fe8b-239a-476a-9213-1eccbd58958e\") " pod="openstack/ovsdbserver-nb-0" Nov 23 06:57:54 crc kubenswrapper[4559]: I1123 06:57:54.198477 4559 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"5d19fe8b-239a-476a-9213-1eccbd58958e\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-nb-0" Nov 23 06:57:54 crc kubenswrapper[4559]: I1123 06:57:54.200041 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5d19fe8b-239a-476a-9213-1eccbd58958e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"5d19fe8b-239a-476a-9213-1eccbd58958e\") " pod="openstack/ovsdbserver-nb-0" Nov 23 06:57:54 crc kubenswrapper[4559]: I1123 06:57:54.201567 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d19fe8b-239a-476a-9213-1eccbd58958e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"5d19fe8b-239a-476a-9213-1eccbd58958e\") " pod="openstack/ovsdbserver-nb-0" Nov 23 06:57:54 crc kubenswrapper[4559]: I1123 06:57:54.206372 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d19fe8b-239a-476a-9213-1eccbd58958e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"5d19fe8b-239a-476a-9213-1eccbd58958e\") " pod="openstack/ovsdbserver-nb-0" Nov 23 06:57:54 crc kubenswrapper[4559]: I1123 06:57:54.206450 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d19fe8b-239a-476a-9213-1eccbd58958e-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5d19fe8b-239a-476a-9213-1eccbd58958e\") " pod="openstack/ovsdbserver-nb-0" Nov 23 06:57:54 crc kubenswrapper[4559]: I1123 06:57:54.207624 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d19fe8b-239a-476a-9213-1eccbd58958e-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5d19fe8b-239a-476a-9213-1eccbd58958e\") " pod="openstack/ovsdbserver-nb-0" Nov 23 06:57:54 crc kubenswrapper[4559]: I1123 06:57:54.214395 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7qmd\" (UniqueName: \"kubernetes.io/projected/5d19fe8b-239a-476a-9213-1eccbd58958e-kube-api-access-g7qmd\") pod \"ovsdbserver-nb-0\" (UID: \"5d19fe8b-239a-476a-9213-1eccbd58958e\") " pod="openstack/ovsdbserver-nb-0" Nov 23 06:57:54 crc kubenswrapper[4559]: I1123 06:57:54.220504 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"5d19fe8b-239a-476a-9213-1eccbd58958e\") " pod="openstack/ovsdbserver-nb-0" Nov 23 06:57:54 crc kubenswrapper[4559]: I1123 06:57:54.305274 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 23 06:57:55 crc kubenswrapper[4559]: I1123 06:57:55.435957 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kb4bw" Nov 23 06:57:55 crc kubenswrapper[4559]: I1123 06:57:55.436026 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kb4bw" Nov 23 06:57:55 crc kubenswrapper[4559]: I1123 06:57:55.494307 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kb4bw" Nov 23 06:57:55 crc kubenswrapper[4559]: I1123 06:57:55.895770 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kb4bw" Nov 23 06:57:55 crc kubenswrapper[4559]: I1123 06:57:55.940460 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kb4bw"] Nov 23 06:57:56 crc kubenswrapper[4559]: I1123 06:57:56.872296 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 23 06:57:56 crc kubenswrapper[4559]: I1123 06:57:56.874005 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 23 06:57:56 crc kubenswrapper[4559]: I1123 06:57:56.879014 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 23 06:57:56 crc kubenswrapper[4559]: I1123 06:57:56.880252 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 23 06:57:56 crc kubenswrapper[4559]: I1123 06:57:56.880302 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-vtmgt" Nov 23 06:57:56 crc kubenswrapper[4559]: I1123 06:57:56.880310 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 23 06:57:56 crc kubenswrapper[4559]: I1123 06:57:56.881740 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 23 06:57:57 crc kubenswrapper[4559]: I1123 06:57:57.046268 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f162538d-e942-4d5e-b2c5-11ba57054cf4-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f162538d-e942-4d5e-b2c5-11ba57054cf4\") " pod="openstack/ovsdbserver-sb-0" Nov 23 06:57:57 crc kubenswrapper[4559]: I1123 06:57:57.046436 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f162538d-e942-4d5e-b2c5-11ba57054cf4-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"f162538d-e942-4d5e-b2c5-11ba57054cf4\") " pod="openstack/ovsdbserver-sb-0" Nov 23 06:57:57 crc kubenswrapper[4559]: I1123 06:57:57.046469 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f162538d-e942-4d5e-b2c5-11ba57054cf4\") " pod="openstack/ovsdbserver-sb-0" Nov 23 06:57:57 crc kubenswrapper[4559]: I1123 06:57:57.046514 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjlfd\" (UniqueName: \"kubernetes.io/projected/f162538d-e942-4d5e-b2c5-11ba57054cf4-kube-api-access-wjlfd\") pod \"ovsdbserver-sb-0\" (UID: \"f162538d-e942-4d5e-b2c5-11ba57054cf4\") " pod="openstack/ovsdbserver-sb-0" Nov 23 06:57:57 crc kubenswrapper[4559]: I1123 06:57:57.046545 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f162538d-e942-4d5e-b2c5-11ba57054cf4-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"f162538d-e942-4d5e-b2c5-11ba57054cf4\") " pod="openstack/ovsdbserver-sb-0" Nov 23 06:57:57 crc kubenswrapper[4559]: I1123 06:57:57.046576 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f162538d-e942-4d5e-b2c5-11ba57054cf4-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"f162538d-e942-4d5e-b2c5-11ba57054cf4\") " pod="openstack/ovsdbserver-sb-0" Nov 23 06:57:57 crc kubenswrapper[4559]: I1123 06:57:57.046605 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f162538d-e942-4d5e-b2c5-11ba57054cf4-config\") pod \"ovsdbserver-sb-0\" (UID: \"f162538d-e942-4d5e-b2c5-11ba57054cf4\") " pod="openstack/ovsdbserver-sb-0" Nov 23 06:57:57 crc kubenswrapper[4559]: I1123 06:57:57.046621 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f162538d-e942-4d5e-b2c5-11ba57054cf4-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f162538d-e942-4d5e-b2c5-11ba57054cf4\") " pod="openstack/ovsdbserver-sb-0" Nov 23 06:57:57 crc kubenswrapper[4559]: I1123 06:57:57.147355 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f162538d-e942-4d5e-b2c5-11ba57054cf4-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"f162538d-e942-4d5e-b2c5-11ba57054cf4\") " pod="openstack/ovsdbserver-sb-0" Nov 23 06:57:57 crc kubenswrapper[4559]: I1123 06:57:57.147399 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f162538d-e942-4d5e-b2c5-11ba57054cf4\") " pod="openstack/ovsdbserver-sb-0" Nov 23 06:57:57 crc kubenswrapper[4559]: I1123 06:57:57.147428 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjlfd\" (UniqueName: \"kubernetes.io/projected/f162538d-e942-4d5e-b2c5-11ba57054cf4-kube-api-access-wjlfd\") pod \"ovsdbserver-sb-0\" (UID: \"f162538d-e942-4d5e-b2c5-11ba57054cf4\") " pod="openstack/ovsdbserver-sb-0" Nov 23 06:57:57 crc kubenswrapper[4559]: I1123 06:57:57.147452 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f162538d-e942-4d5e-b2c5-11ba57054cf4-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"f162538d-e942-4d5e-b2c5-11ba57054cf4\") " pod="openstack/ovsdbserver-sb-0" Nov 23 06:57:57 crc kubenswrapper[4559]: I1123 06:57:57.147466 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f162538d-e942-4d5e-b2c5-11ba57054cf4-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"f162538d-e942-4d5e-b2c5-11ba57054cf4\") " pod="openstack/ovsdbserver-sb-0" Nov 23 06:57:57 crc kubenswrapper[4559]: I1123 06:57:57.147497 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f162538d-e942-4d5e-b2c5-11ba57054cf4-config\") pod \"ovsdbserver-sb-0\" (UID: \"f162538d-e942-4d5e-b2c5-11ba57054cf4\") " pod="openstack/ovsdbserver-sb-0" Nov 23 06:57:57 crc kubenswrapper[4559]: I1123 06:57:57.147510 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f162538d-e942-4d5e-b2c5-11ba57054cf4-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f162538d-e942-4d5e-b2c5-11ba57054cf4\") " pod="openstack/ovsdbserver-sb-0" Nov 23 06:57:57 crc kubenswrapper[4559]: I1123 06:57:57.147542 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f162538d-e942-4d5e-b2c5-11ba57054cf4-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f162538d-e942-4d5e-b2c5-11ba57054cf4\") " pod="openstack/ovsdbserver-sb-0" Nov 23 06:57:57 crc kubenswrapper[4559]: I1123 06:57:57.148075 4559 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f162538d-e942-4d5e-b2c5-11ba57054cf4\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-sb-0" Nov 23 06:57:57 crc kubenswrapper[4559]: I1123 06:57:57.148494 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f162538d-e942-4d5e-b2c5-11ba57054cf4-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"f162538d-e942-4d5e-b2c5-11ba57054cf4\") " pod="openstack/ovsdbserver-sb-0" Nov 23 06:57:57 crc kubenswrapper[4559]: I1123 06:57:57.148956 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f162538d-e942-4d5e-b2c5-11ba57054cf4-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"f162538d-e942-4d5e-b2c5-11ba57054cf4\") " pod="openstack/ovsdbserver-sb-0" Nov 23 06:57:57 crc kubenswrapper[4559]: I1123 06:57:57.150198 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f162538d-e942-4d5e-b2c5-11ba57054cf4-config\") pod \"ovsdbserver-sb-0\" (UID: \"f162538d-e942-4d5e-b2c5-11ba57054cf4\") " pod="openstack/ovsdbserver-sb-0" Nov 23 06:57:57 crc kubenswrapper[4559]: I1123 06:57:57.154317 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f162538d-e942-4d5e-b2c5-11ba57054cf4-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f162538d-e942-4d5e-b2c5-11ba57054cf4\") " pod="openstack/ovsdbserver-sb-0" Nov 23 06:57:57 crc kubenswrapper[4559]: I1123 06:57:57.155091 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f162538d-e942-4d5e-b2c5-11ba57054cf4-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f162538d-e942-4d5e-b2c5-11ba57054cf4\") " pod="openstack/ovsdbserver-sb-0" Nov 23 06:57:57 crc kubenswrapper[4559]: I1123 06:57:57.158904 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f162538d-e942-4d5e-b2c5-11ba57054cf4-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"f162538d-e942-4d5e-b2c5-11ba57054cf4\") " pod="openstack/ovsdbserver-sb-0" Nov 23 06:57:57 crc kubenswrapper[4559]: I1123 06:57:57.169576 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f162538d-e942-4d5e-b2c5-11ba57054cf4\") " pod="openstack/ovsdbserver-sb-0" Nov 23 06:57:57 crc kubenswrapper[4559]: I1123 06:57:57.171880 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjlfd\" (UniqueName: \"kubernetes.io/projected/f162538d-e942-4d5e-b2c5-11ba57054cf4-kube-api-access-wjlfd\") pod \"ovsdbserver-sb-0\" (UID: \"f162538d-e942-4d5e-b2c5-11ba57054cf4\") " pod="openstack/ovsdbserver-sb-0" Nov 23 06:57:57 crc kubenswrapper[4559]: I1123 06:57:57.208279 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 23 06:57:57 crc kubenswrapper[4559]: I1123 06:57:57.821613 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kb4bw" podUID="d6f08db9-323e-43dd-b925-1eea5dd88345" containerName="registry-server" containerID="cri-o://39e183bd477c83b9156f9160c9674fa9fb44501acbb4b1d1ae22fe79ecce18d1" gracePeriod=2 Nov 23 06:57:58 crc kubenswrapper[4559]: I1123 06:57:58.830492 4559 generic.go:334] "Generic (PLEG): container finished" podID="d6f08db9-323e-43dd-b925-1eea5dd88345" containerID="39e183bd477c83b9156f9160c9674fa9fb44501acbb4b1d1ae22fe79ecce18d1" exitCode=0 Nov 23 06:57:58 crc kubenswrapper[4559]: I1123 06:57:58.830558 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kb4bw" event={"ID":"d6f08db9-323e-43dd-b925-1eea5dd88345","Type":"ContainerDied","Data":"39e183bd477c83b9156f9160c9674fa9fb44501acbb4b1d1ae22fe79ecce18d1"} Nov 23 06:57:59 crc kubenswrapper[4559]: I1123 06:57:59.498265 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 23 06:58:05 crc kubenswrapper[4559]: E1123 06:58:05.436535 4559 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 39e183bd477c83b9156f9160c9674fa9fb44501acbb4b1d1ae22fe79ecce18d1 is running failed: container process not found" containerID="39e183bd477c83b9156f9160c9674fa9fb44501acbb4b1d1ae22fe79ecce18d1" cmd=["grpc_health_probe","-addr=:50051"] Nov 23 06:58:05 crc kubenswrapper[4559]: E1123 06:58:05.437520 4559 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 39e183bd477c83b9156f9160c9674fa9fb44501acbb4b1d1ae22fe79ecce18d1 is running failed: container process not found" containerID="39e183bd477c83b9156f9160c9674fa9fb44501acbb4b1d1ae22fe79ecce18d1" cmd=["grpc_health_probe","-addr=:50051"] Nov 23 06:58:05 crc kubenswrapper[4559]: E1123 06:58:05.437892 4559 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 39e183bd477c83b9156f9160c9674fa9fb44501acbb4b1d1ae22fe79ecce18d1 is running failed: container process not found" containerID="39e183bd477c83b9156f9160c9674fa9fb44501acbb4b1d1ae22fe79ecce18d1" cmd=["grpc_health_probe","-addr=:50051"] Nov 23 06:58:05 crc kubenswrapper[4559]: E1123 06:58:05.437933 4559 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 39e183bd477c83b9156f9160c9674fa9fb44501acbb4b1d1ae22fe79ecce18d1 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-kb4bw" podUID="d6f08db9-323e-43dd-b925-1eea5dd88345" containerName="registry-server" Nov 23 06:58:08 crc kubenswrapper[4559]: E1123 06:58:08.683604 4559 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:8e43c662a6abf8c9a07ada252f8dc6af" Nov 23 06:58:08 crc kubenswrapper[4559]: E1123 06:58:08.683914 4559 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:8e43c662a6abf8c9a07ada252f8dc6af" Nov 23 06:58:08 crc kubenswrapper[4559]: E1123 06:58:08.684069 4559 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:8e43c662a6abf8c9a07ada252f8dc6af,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s9qg9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(61e2f87a-db68-432a-8d13-94087d1afb17): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 06:58:08 crc kubenswrapper[4559]: E1123 06:58:08.685249 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="61e2f87a-db68-432a-8d13-94087d1afb17" Nov 23 06:58:08 crc kubenswrapper[4559]: E1123 06:58:08.692737 4559 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:8e43c662a6abf8c9a07ada252f8dc6af" Nov 23 06:58:08 crc kubenswrapper[4559]: E1123 06:58:08.692783 4559 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:8e43c662a6abf8c9a07ada252f8dc6af" Nov 23 06:58:08 crc kubenswrapper[4559]: E1123 06:58:08.692924 4559 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:8e43c662a6abf8c9a07ada252f8dc6af,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zd2cg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(a0fbc3f6-1c87-495a-a38f-5a0ae768ee20): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 06:58:08 crc kubenswrapper[4559]: E1123 06:58:08.694140 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="a0fbc3f6-1c87-495a-a38f-5a0ae768ee20" Nov 23 06:58:08 crc kubenswrapper[4559]: E1123 06:58:08.913891 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:8e43c662a6abf8c9a07ada252f8dc6af\\\"\"" pod="openstack/rabbitmq-server-0" podUID="a0fbc3f6-1c87-495a-a38f-5a0ae768ee20" Nov 23 06:58:08 crc kubenswrapper[4559]: E1123 06:58:08.914430 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:8e43c662a6abf8c9a07ada252f8dc6af\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="61e2f87a-db68-432a-8d13-94087d1afb17" Nov 23 06:58:10 crc kubenswrapper[4559]: I1123 06:58:10.123222 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kb4bw" Nov 23 06:58:10 crc kubenswrapper[4559]: I1123 06:58:10.276470 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99psz\" (UniqueName: \"kubernetes.io/projected/d6f08db9-323e-43dd-b925-1eea5dd88345-kube-api-access-99psz\") pod \"d6f08db9-323e-43dd-b925-1eea5dd88345\" (UID: \"d6f08db9-323e-43dd-b925-1eea5dd88345\") " Nov 23 06:58:10 crc kubenswrapper[4559]: I1123 06:58:10.276695 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6f08db9-323e-43dd-b925-1eea5dd88345-utilities\") pod \"d6f08db9-323e-43dd-b925-1eea5dd88345\" (UID: \"d6f08db9-323e-43dd-b925-1eea5dd88345\") " Nov 23 06:58:10 crc kubenswrapper[4559]: I1123 06:58:10.276783 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6f08db9-323e-43dd-b925-1eea5dd88345-catalog-content\") pod \"d6f08db9-323e-43dd-b925-1eea5dd88345\" (UID: \"d6f08db9-323e-43dd-b925-1eea5dd88345\") " Nov 23 06:58:10 crc kubenswrapper[4559]: I1123 06:58:10.278057 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6f08db9-323e-43dd-b925-1eea5dd88345-utilities" (OuterVolumeSpecName: "utilities") pod "d6f08db9-323e-43dd-b925-1eea5dd88345" (UID: "d6f08db9-323e-43dd-b925-1eea5dd88345"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:58:10 crc kubenswrapper[4559]: I1123 06:58:10.279229 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6f08db9-323e-43dd-b925-1eea5dd88345-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:10 crc kubenswrapper[4559]: I1123 06:58:10.299089 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6f08db9-323e-43dd-b925-1eea5dd88345-kube-api-access-99psz" (OuterVolumeSpecName: "kube-api-access-99psz") pod "d6f08db9-323e-43dd-b925-1eea5dd88345" (UID: "d6f08db9-323e-43dd-b925-1eea5dd88345"). InnerVolumeSpecName "kube-api-access-99psz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:58:10 crc kubenswrapper[4559]: I1123 06:58:10.337048 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6f08db9-323e-43dd-b925-1eea5dd88345-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d6f08db9-323e-43dd-b925-1eea5dd88345" (UID: "d6f08db9-323e-43dd-b925-1eea5dd88345"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:58:10 crc kubenswrapper[4559]: I1123 06:58:10.380813 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99psz\" (UniqueName: \"kubernetes.io/projected/d6f08db9-323e-43dd-b925-1eea5dd88345-kube-api-access-99psz\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:10 crc kubenswrapper[4559]: I1123 06:58:10.380851 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6f08db9-323e-43dd-b925-1eea5dd88345-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:10 crc kubenswrapper[4559]: I1123 06:58:10.936188 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kb4bw" event={"ID":"d6f08db9-323e-43dd-b925-1eea5dd88345","Type":"ContainerDied","Data":"7a39a4b1f615be0f2a41503eef627e7e7b74d67967748ec20f08cb31635d27e4"} Nov 23 06:58:10 crc kubenswrapper[4559]: I1123 06:58:10.936980 4559 scope.go:117] "RemoveContainer" containerID="39e183bd477c83b9156f9160c9674fa9fb44501acbb4b1d1ae22fe79ecce18d1" Nov 23 06:58:10 crc kubenswrapper[4559]: I1123 06:58:10.936265 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kb4bw" Nov 23 06:58:10 crc kubenswrapper[4559]: I1123 06:58:10.972202 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kb4bw"] Nov 23 06:58:10 crc kubenswrapper[4559]: I1123 06:58:10.977530 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kb4bw"] Nov 23 06:58:11 crc kubenswrapper[4559]: E1123 06:58:11.212615 4559 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-memcached:8e43c662a6abf8c9a07ada252f8dc6af" Nov 23 06:58:11 crc kubenswrapper[4559]: E1123 06:58:11.212724 4559 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-memcached:8e43c662a6abf8c9a07ada252f8dc6af" Nov 23 06:58:11 crc kubenswrapper[4559]: E1123 06:58:11.212915 4559 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-memcached:8e43c662a6abf8c9a07ada252f8dc6af,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n5c9hd6h689h5dch8dh5bfh66ch68bh54ch585hdfh8dh67bhb4h5b5h65fh564h546h5fdhb8h54bh644h65bh575h56dh6dh5dbh8ch9h587h6hf4q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b6bqj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(de0cb0c4-ffde-44c5-adb9-7aea0692c1b5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 06:58:11 crc kubenswrapper[4559]: E1123 06:58:11.214210 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="de0cb0c4-ffde-44c5-adb9-7aea0692c1b5" Nov 23 06:58:11 crc kubenswrapper[4559]: E1123 06:58:11.252216 4559 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-neutron-server:8e43c662a6abf8c9a07ada252f8dc6af" Nov 23 06:58:11 crc kubenswrapper[4559]: E1123 06:58:11.252281 4559 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-neutron-server:8e43c662a6abf8c9a07ada252f8dc6af" Nov 23 06:58:11 crc kubenswrapper[4559]: E1123 06:58:11.252480 4559 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-neutron-server:8e43c662a6abf8c9a07ada252f8dc6af,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d2zrd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-84bd59c769-rdgpv_openstack(19172889-a652-4d9e-a074-1dc45485752f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 06:58:11 crc kubenswrapper[4559]: E1123 06:58:11.253743 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-84bd59c769-rdgpv" podUID="19172889-a652-4d9e-a074-1dc45485752f" Nov 23 06:58:11 crc kubenswrapper[4559]: I1123 06:58:11.259595 4559 scope.go:117] "RemoveContainer" containerID="0a77d5d8f83d5f50554f687a003de50e674d73f27e63c9d6af603da4f236b674" Nov 23 06:58:11 crc kubenswrapper[4559]: I1123 06:58:11.360100 4559 scope.go:117] "RemoveContainer" containerID="c12d69435ac9c13967e865a1980550ffe3c0f8e3660ba9b21a9f513ff88b6c14" Nov 23 06:58:11 crc kubenswrapper[4559]: I1123 06:58:11.742479 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 23 06:58:11 crc kubenswrapper[4559]: I1123 06:58:11.754528 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-blj85"] Nov 23 06:58:11 crc kubenswrapper[4559]: W1123 06:58:11.854806 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd534f227_972b_4a85_aeb5_8f1d226352e1.slice/crio-9b971dd1c7df94d17011d71765a28a7937f9afdf435e56d35fa4d6bc3953e5f8 WatchSource:0}: Error finding container 9b971dd1c7df94d17011d71765a28a7937f9afdf435e56d35fa4d6bc3953e5f8: Status 404 returned error can't find the container with id 9b971dd1c7df94d17011d71765a28a7937f9afdf435e56d35fa4d6bc3953e5f8 Nov 23 06:58:11 crc kubenswrapper[4559]: I1123 06:58:11.856500 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-jg7hk"] Nov 23 06:58:11 crc kubenswrapper[4559]: I1123 06:58:11.947611 4559 generic.go:334] "Generic (PLEG): container finished" podID="ebba232a-99bb-4f45-a0ce-3628f42dcc77" containerID="b21e2437402d5251e7379fd98e040e46e24f6b535fb2e9369a5171dfdc104d67" exitCode=0 Nov 23 06:58:11 crc kubenswrapper[4559]: I1123 06:58:11.947763 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f47fdfb89-xjlhn" event={"ID":"ebba232a-99bb-4f45-a0ce-3628f42dcc77","Type":"ContainerDied","Data":"b21e2437402d5251e7379fd98e040e46e24f6b535fb2e9369a5171dfdc104d67"} Nov 23 06:58:11 crc kubenswrapper[4559]: I1123 06:58:11.950255 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"01dff7d1-bf4c-4d88-900c-99de62949ac7","Type":"ContainerStarted","Data":"5672161e23aaf5562043f18d480168d5f4970fd9b172228537b53d52c1ac3cff"} Nov 23 06:58:11 crc kubenswrapper[4559]: I1123 06:58:11.952812 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"5d19fe8b-239a-476a-9213-1eccbd58958e","Type":"ContainerStarted","Data":"e65fff358b9d25be637a35ef3178d54e4c39a0ebd028efd8d3a17d3bef1fe103"} Nov 23 06:58:11 crc kubenswrapper[4559]: I1123 06:58:11.955043 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"e097ab60-4653-4038-b6de-40b55936565c","Type":"ContainerStarted","Data":"22979e4cf1a2c4da38a958c3d8455c8cbfa657871fcc079f0aea832c1e1d9b99"} Nov 23 06:58:11 crc kubenswrapper[4559]: I1123 06:58:11.957711 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-blj85" event={"ID":"20662e83-ba79-4c97-80fb-98fda28c1149","Type":"ContainerStarted","Data":"bce457967547bf4b7725922a92a21e848a1f855887a35764c21a12520da09c6e"} Nov 23 06:58:11 crc kubenswrapper[4559]: I1123 06:58:11.959174 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jg7hk" event={"ID":"d534f227-972b-4a85-aeb5-8f1d226352e1","Type":"ContainerStarted","Data":"9b971dd1c7df94d17011d71765a28a7937f9afdf435e56d35fa4d6bc3953e5f8"} Nov 23 06:58:11 crc kubenswrapper[4559]: I1123 06:58:11.961061 4559 generic.go:334] "Generic (PLEG): container finished" podID="0e6213f8-9948-47e9-88ac-7d7bdaddac26" containerID="ff867a700a523808a42c0dc55d10a2a7c544c16dc2e540c82cc4bdd7eeaae817" exitCode=0 Nov 23 06:58:11 crc kubenswrapper[4559]: I1123 06:58:11.961143 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-759c6cc4df-s47tx" event={"ID":"0e6213f8-9948-47e9-88ac-7d7bdaddac26","Type":"ContainerDied","Data":"ff867a700a523808a42c0dc55d10a2a7c544c16dc2e540c82cc4bdd7eeaae817"} Nov 23 06:58:11 crc kubenswrapper[4559]: I1123 06:58:11.963594 4559 generic.go:334] "Generic (PLEG): container finished" podID="9caa7f7e-01ab-4998-82bd-df9380117fb9" containerID="bbfb23a0059243f6832b7c3bfb089a93e1ae59274646b67219cf2b571e705cad" exitCode=0 Nov 23 06:58:11 crc kubenswrapper[4559]: I1123 06:58:11.963689 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6794664cc7-vmnd9" event={"ID":"9caa7f7e-01ab-4998-82bd-df9380117fb9","Type":"ContainerDied","Data":"bbfb23a0059243f6832b7c3bfb089a93e1ae59274646b67219cf2b571e705cad"} Nov 23 06:58:11 crc kubenswrapper[4559]: E1123 06:58:11.967574 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-memcached:8e43c662a6abf8c9a07ada252f8dc6af\\\"\"" pod="openstack/memcached-0" podUID="de0cb0c4-ffde-44c5-adb9-7aea0692c1b5" Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.292147 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6f08db9-323e-43dd-b925-1eea5dd88345" path="/var/lib/kubelet/pods/d6f08db9-323e-43dd-b925-1eea5dd88345/volumes" Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.357769 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6794664cc7-vmnd9" Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.361900 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84bd59c769-rdgpv" Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.521950 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19172889-a652-4d9e-a074-1dc45485752f-dns-svc\") pod \"19172889-a652-4d9e-a074-1dc45485752f\" (UID: \"19172889-a652-4d9e-a074-1dc45485752f\") " Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.522050 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19172889-a652-4d9e-a074-1dc45485752f-config\") pod \"19172889-a652-4d9e-a074-1dc45485752f\" (UID: \"19172889-a652-4d9e-a074-1dc45485752f\") " Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.522074 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9caa7f7e-01ab-4998-82bd-df9380117fb9-config\") pod \"9caa7f7e-01ab-4998-82bd-df9380117fb9\" (UID: \"9caa7f7e-01ab-4998-82bd-df9380117fb9\") " Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.522105 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2zrd\" (UniqueName: \"kubernetes.io/projected/19172889-a652-4d9e-a074-1dc45485752f-kube-api-access-d2zrd\") pod \"19172889-a652-4d9e-a074-1dc45485752f\" (UID: \"19172889-a652-4d9e-a074-1dc45485752f\") " Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.522163 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q64zk\" (UniqueName: \"kubernetes.io/projected/9caa7f7e-01ab-4998-82bd-df9380117fb9-kube-api-access-q64zk\") pod \"9caa7f7e-01ab-4998-82bd-df9380117fb9\" (UID: \"9caa7f7e-01ab-4998-82bd-df9380117fb9\") " Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.522482 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19172889-a652-4d9e-a074-1dc45485752f-config" (OuterVolumeSpecName: "config") pod "19172889-a652-4d9e-a074-1dc45485752f" (UID: "19172889-a652-4d9e-a074-1dc45485752f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.522629 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19172889-a652-4d9e-a074-1dc45485752f-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.522622 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19172889-a652-4d9e-a074-1dc45485752f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "19172889-a652-4d9e-a074-1dc45485752f" (UID: "19172889-a652-4d9e-a074-1dc45485752f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.527435 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19172889-a652-4d9e-a074-1dc45485752f-kube-api-access-d2zrd" (OuterVolumeSpecName: "kube-api-access-d2zrd") pod "19172889-a652-4d9e-a074-1dc45485752f" (UID: "19172889-a652-4d9e-a074-1dc45485752f"). InnerVolumeSpecName "kube-api-access-d2zrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.527530 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9caa7f7e-01ab-4998-82bd-df9380117fb9-kube-api-access-q64zk" (OuterVolumeSpecName: "kube-api-access-q64zk") pod "9caa7f7e-01ab-4998-82bd-df9380117fb9" (UID: "9caa7f7e-01ab-4998-82bd-df9380117fb9"). InnerVolumeSpecName "kube-api-access-q64zk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.539211 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9caa7f7e-01ab-4998-82bd-df9380117fb9-config" (OuterVolumeSpecName: "config") pod "9caa7f7e-01ab-4998-82bd-df9380117fb9" (UID: "9caa7f7e-01ab-4998-82bd-df9380117fb9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.623999 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9caa7f7e-01ab-4998-82bd-df9380117fb9-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.624029 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2zrd\" (UniqueName: \"kubernetes.io/projected/19172889-a652-4d9e-a074-1dc45485752f-kube-api-access-d2zrd\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.624042 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q64zk\" (UniqueName: \"kubernetes.io/projected/9caa7f7e-01ab-4998-82bd-df9380117fb9-kube-api-access-q64zk\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.624053 4559 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19172889-a652-4d9e-a074-1dc45485752f-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.724128 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 23 06:58:12 crc kubenswrapper[4559]: W1123 06:58:12.732711 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf162538d_e942_4d5e_b2c5_11ba57054cf4.slice/crio-4df05a7c03e879cae846f1976d3f5f78128b0342b314fd5d8ebfc018a3178de8 WatchSource:0}: Error finding container 4df05a7c03e879cae846f1976d3f5f78128b0342b314fd5d8ebfc018a3178de8: Status 404 returned error can't find the container with id 4df05a7c03e879cae846f1976d3f5f78128b0342b314fd5d8ebfc018a3178de8 Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.980477 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84bd59c769-rdgpv" Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.980470 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84bd59c769-rdgpv" event={"ID":"19172889-a652-4d9e-a074-1dc45485752f","Type":"ContainerDied","Data":"05137c958571eb560e0bc8eb13fcb49afbce22fe98cf86ed0784f9f868a8a21e"} Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.983952 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f47fdfb89-xjlhn" event={"ID":"ebba232a-99bb-4f45-a0ce-3628f42dcc77","Type":"ContainerStarted","Data":"3e21393020565de7f1e1fe3808086932a925cb97580ccf93752fd78328a9da20"} Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.984144 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7f47fdfb89-xjlhn" Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.989502 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-759c6cc4df-s47tx" event={"ID":"0e6213f8-9948-47e9-88ac-7d7bdaddac26","Type":"ContainerStarted","Data":"49c281f5966b944fb67c9b11ed9bf54b265407038016add98feccbd852cc8661"} Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.989753 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-759c6cc4df-s47tx" Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.992894 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6794664cc7-vmnd9" event={"ID":"9caa7f7e-01ab-4998-82bd-df9380117fb9","Type":"ContainerDied","Data":"47bb93c7efe950640763715beffa55bf04af109f5eb4d8447badc01b6670cc24"} Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.992939 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6794664cc7-vmnd9" Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.992966 4559 scope.go:117] "RemoveContainer" containerID="bbfb23a0059243f6832b7c3bfb089a93e1ae59274646b67219cf2b571e705cad" Nov 23 06:58:12 crc kubenswrapper[4559]: I1123 06:58:12.994419 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f162538d-e942-4d5e-b2c5-11ba57054cf4","Type":"ContainerStarted","Data":"4df05a7c03e879cae846f1976d3f5f78128b0342b314fd5d8ebfc018a3178de8"} Nov 23 06:58:13 crc kubenswrapper[4559]: I1123 06:58:13.005025 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7f47fdfb89-xjlhn" podStartSLOduration=3.28293511 podStartE2EDuration="31.005010809s" podCreationTimestamp="2025-11-23 06:57:42 +0000 UTC" firstStartedPulling="2025-11-23 06:57:43.652651198 +0000 UTC m=+765.674636812" lastFinishedPulling="2025-11-23 06:58:11.374726897 +0000 UTC m=+793.396712511" observedRunningTime="2025-11-23 06:58:13.003108752 +0000 UTC m=+795.025094367" watchObservedRunningTime="2025-11-23 06:58:13.005010809 +0000 UTC m=+795.026996423" Nov 23 06:58:13 crc kubenswrapper[4559]: I1123 06:58:13.025172 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-759c6cc4df-s47tx" podStartSLOduration=2.503707502 podStartE2EDuration="30.025161531s" podCreationTimestamp="2025-11-23 06:57:43 +0000 UTC" firstStartedPulling="2025-11-23 06:57:43.844401415 +0000 UTC m=+765.866387030" lastFinishedPulling="2025-11-23 06:58:11.365855445 +0000 UTC m=+793.387841059" observedRunningTime="2025-11-23 06:58:13.019430935 +0000 UTC m=+795.041416549" watchObservedRunningTime="2025-11-23 06:58:13.025161531 +0000 UTC m=+795.047147144" Nov 23 06:58:13 crc kubenswrapper[4559]: I1123 06:58:13.056195 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84bd59c769-rdgpv"] Nov 23 06:58:13 crc kubenswrapper[4559]: I1123 06:58:13.067169 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84bd59c769-rdgpv"] Nov 23 06:58:13 crc kubenswrapper[4559]: I1123 06:58:13.080389 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6794664cc7-vmnd9"] Nov 23 06:58:13 crc kubenswrapper[4559]: I1123 06:58:13.084359 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6794664cc7-vmnd9"] Nov 23 06:58:14 crc kubenswrapper[4559]: I1123 06:58:14.286239 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19172889-a652-4d9e-a074-1dc45485752f" path="/var/lib/kubelet/pods/19172889-a652-4d9e-a074-1dc45485752f/volumes" Nov 23 06:58:14 crc kubenswrapper[4559]: I1123 06:58:14.287250 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9caa7f7e-01ab-4998-82bd-df9380117fb9" path="/var/lib/kubelet/pods/9caa7f7e-01ab-4998-82bd-df9380117fb9/volumes" Nov 23 06:58:14 crc kubenswrapper[4559]: I1123 06:58:14.741623 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9v64g"] Nov 23 06:58:14 crc kubenswrapper[4559]: E1123 06:58:14.743148 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6f08db9-323e-43dd-b925-1eea5dd88345" containerName="extract-content" Nov 23 06:58:14 crc kubenswrapper[4559]: I1123 06:58:14.743163 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6f08db9-323e-43dd-b925-1eea5dd88345" containerName="extract-content" Nov 23 06:58:14 crc kubenswrapper[4559]: E1123 06:58:14.743175 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6f08db9-323e-43dd-b925-1eea5dd88345" containerName="registry-server" Nov 23 06:58:14 crc kubenswrapper[4559]: I1123 06:58:14.743181 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6f08db9-323e-43dd-b925-1eea5dd88345" containerName="registry-server" Nov 23 06:58:14 crc kubenswrapper[4559]: E1123 06:58:14.743220 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6f08db9-323e-43dd-b925-1eea5dd88345" containerName="extract-utilities" Nov 23 06:58:14 crc kubenswrapper[4559]: I1123 06:58:14.743226 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6f08db9-323e-43dd-b925-1eea5dd88345" containerName="extract-utilities" Nov 23 06:58:14 crc kubenswrapper[4559]: E1123 06:58:14.743236 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9caa7f7e-01ab-4998-82bd-df9380117fb9" containerName="init" Nov 23 06:58:14 crc kubenswrapper[4559]: I1123 06:58:14.743242 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="9caa7f7e-01ab-4998-82bd-df9380117fb9" containerName="init" Nov 23 06:58:14 crc kubenswrapper[4559]: I1123 06:58:14.743467 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6f08db9-323e-43dd-b925-1eea5dd88345" containerName="registry-server" Nov 23 06:58:14 crc kubenswrapper[4559]: I1123 06:58:14.743487 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="9caa7f7e-01ab-4998-82bd-df9380117fb9" containerName="init" Nov 23 06:58:14 crc kubenswrapper[4559]: I1123 06:58:14.745159 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9v64g" Nov 23 06:58:14 crc kubenswrapper[4559]: I1123 06:58:14.747563 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9v64g"] Nov 23 06:58:14 crc kubenswrapper[4559]: I1123 06:58:14.867596 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqm4q\" (UniqueName: \"kubernetes.io/projected/56c603bf-2cb8-4c83-b11f-7794dcc0114a-kube-api-access-rqm4q\") pod \"community-operators-9v64g\" (UID: \"56c603bf-2cb8-4c83-b11f-7794dcc0114a\") " pod="openshift-marketplace/community-operators-9v64g" Nov 23 06:58:14 crc kubenswrapper[4559]: I1123 06:58:14.868114 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56c603bf-2cb8-4c83-b11f-7794dcc0114a-catalog-content\") pod \"community-operators-9v64g\" (UID: \"56c603bf-2cb8-4c83-b11f-7794dcc0114a\") " pod="openshift-marketplace/community-operators-9v64g" Nov 23 06:58:14 crc kubenswrapper[4559]: I1123 06:58:14.868244 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56c603bf-2cb8-4c83-b11f-7794dcc0114a-utilities\") pod \"community-operators-9v64g\" (UID: \"56c603bf-2cb8-4c83-b11f-7794dcc0114a\") " pod="openshift-marketplace/community-operators-9v64g" Nov 23 06:58:14 crc kubenswrapper[4559]: I1123 06:58:14.969731 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqm4q\" (UniqueName: \"kubernetes.io/projected/56c603bf-2cb8-4c83-b11f-7794dcc0114a-kube-api-access-rqm4q\") pod \"community-operators-9v64g\" (UID: \"56c603bf-2cb8-4c83-b11f-7794dcc0114a\") " pod="openshift-marketplace/community-operators-9v64g" Nov 23 06:58:14 crc kubenswrapper[4559]: I1123 06:58:14.969802 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56c603bf-2cb8-4c83-b11f-7794dcc0114a-catalog-content\") pod \"community-operators-9v64g\" (UID: \"56c603bf-2cb8-4c83-b11f-7794dcc0114a\") " pod="openshift-marketplace/community-operators-9v64g" Nov 23 06:58:14 crc kubenswrapper[4559]: I1123 06:58:14.969857 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56c603bf-2cb8-4c83-b11f-7794dcc0114a-utilities\") pod \"community-operators-9v64g\" (UID: \"56c603bf-2cb8-4c83-b11f-7794dcc0114a\") " pod="openshift-marketplace/community-operators-9v64g" Nov 23 06:58:14 crc kubenswrapper[4559]: I1123 06:58:14.970306 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56c603bf-2cb8-4c83-b11f-7794dcc0114a-utilities\") pod \"community-operators-9v64g\" (UID: \"56c603bf-2cb8-4c83-b11f-7794dcc0114a\") " pod="openshift-marketplace/community-operators-9v64g" Nov 23 06:58:14 crc kubenswrapper[4559]: I1123 06:58:14.970724 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56c603bf-2cb8-4c83-b11f-7794dcc0114a-catalog-content\") pod \"community-operators-9v64g\" (UID: \"56c603bf-2cb8-4c83-b11f-7794dcc0114a\") " pod="openshift-marketplace/community-operators-9v64g" Nov 23 06:58:14 crc kubenswrapper[4559]: I1123 06:58:14.991604 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqm4q\" (UniqueName: \"kubernetes.io/projected/56c603bf-2cb8-4c83-b11f-7794dcc0114a-kube-api-access-rqm4q\") pod \"community-operators-9v64g\" (UID: \"56c603bf-2cb8-4c83-b11f-7794dcc0114a\") " pod="openshift-marketplace/community-operators-9v64g" Nov 23 06:58:15 crc kubenswrapper[4559]: I1123 06:58:15.029483 4559 generic.go:334] "Generic (PLEG): container finished" podID="01dff7d1-bf4c-4d88-900c-99de62949ac7" containerID="5672161e23aaf5562043f18d480168d5f4970fd9b172228537b53d52c1ac3cff" exitCode=0 Nov 23 06:58:15 crc kubenswrapper[4559]: I1123 06:58:15.029538 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"01dff7d1-bf4c-4d88-900c-99de62949ac7","Type":"ContainerDied","Data":"5672161e23aaf5562043f18d480168d5f4970fd9b172228537b53d52c1ac3cff"} Nov 23 06:58:15 crc kubenswrapper[4559]: I1123 06:58:15.035010 4559 generic.go:334] "Generic (PLEG): container finished" podID="e097ab60-4653-4038-b6de-40b55936565c" containerID="22979e4cf1a2c4da38a958c3d8455c8cbfa657871fcc079f0aea832c1e1d9b99" exitCode=0 Nov 23 06:58:15 crc kubenswrapper[4559]: I1123 06:58:15.035061 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"e097ab60-4653-4038-b6de-40b55936565c","Type":"ContainerDied","Data":"22979e4cf1a2c4da38a958c3d8455c8cbfa657871fcc079f0aea832c1e1d9b99"} Nov 23 06:58:15 crc kubenswrapper[4559]: I1123 06:58:15.067697 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9v64g" Nov 23 06:58:15 crc kubenswrapper[4559]: I1123 06:58:15.574792 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9v64g"] Nov 23 06:58:15 crc kubenswrapper[4559]: W1123 06:58:15.584313 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56c603bf_2cb8_4c83_b11f_7794dcc0114a.slice/crio-a80fe4b5bbc3a21c49d0738d4bc88c06988a148d17f4fe09d7b2bf5d1d9b6d13 WatchSource:0}: Error finding container a80fe4b5bbc3a21c49d0738d4bc88c06988a148d17f4fe09d7b2bf5d1d9b6d13: Status 404 returned error can't find the container with id a80fe4b5bbc3a21c49d0738d4bc88c06988a148d17f4fe09d7b2bf5d1d9b6d13 Nov 23 06:58:16 crc kubenswrapper[4559]: I1123 06:58:16.047887 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"01dff7d1-bf4c-4d88-900c-99de62949ac7","Type":"ContainerStarted","Data":"9cb5032ac1f14fae32074c20d4c425e2f19a13d69e7323e04da49ab5fda62a0b"} Nov 23 06:58:16 crc kubenswrapper[4559]: I1123 06:58:16.049555 4559 generic.go:334] "Generic (PLEG): container finished" podID="56c603bf-2cb8-4c83-b11f-7794dcc0114a" containerID="481c1c0cdb7c672d4a4a53468a8b1360b3efaffa13d3dc3b3b8a200072b6173d" exitCode=0 Nov 23 06:58:16 crc kubenswrapper[4559]: I1123 06:58:16.049757 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9v64g" event={"ID":"56c603bf-2cb8-4c83-b11f-7794dcc0114a","Type":"ContainerDied","Data":"481c1c0cdb7c672d4a4a53468a8b1360b3efaffa13d3dc3b3b8a200072b6173d"} Nov 23 06:58:16 crc kubenswrapper[4559]: I1123 06:58:16.050062 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9v64g" event={"ID":"56c603bf-2cb8-4c83-b11f-7794dcc0114a","Type":"ContainerStarted","Data":"a80fe4b5bbc3a21c49d0738d4bc88c06988a148d17f4fe09d7b2bf5d1d9b6d13"} Nov 23 06:58:16 crc kubenswrapper[4559]: I1123 06:58:16.052375 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"e097ab60-4653-4038-b6de-40b55936565c","Type":"ContainerStarted","Data":"29ccbb7dddb9b38d384a92a45d9068499ecfc23d7225fbee0d0857ace55179fc"} Nov 23 06:58:16 crc kubenswrapper[4559]: I1123 06:58:16.068931 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=6.913326023 podStartE2EDuration="30.068919012s" podCreationTimestamp="2025-11-23 06:57:46 +0000 UTC" firstStartedPulling="2025-11-23 06:57:48.208753849 +0000 UTC m=+770.230739463" lastFinishedPulling="2025-11-23 06:58:11.364346838 +0000 UTC m=+793.386332452" observedRunningTime="2025-11-23 06:58:16.068351285 +0000 UTC m=+798.090336899" watchObservedRunningTime="2025-11-23 06:58:16.068919012 +0000 UTC m=+798.090904626" Nov 23 06:58:16 crc kubenswrapper[4559]: I1123 06:58:16.088992 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=7.365036336 podStartE2EDuration="32.088973053s" podCreationTimestamp="2025-11-23 06:57:44 +0000 UTC" firstStartedPulling="2025-11-23 06:57:46.545360398 +0000 UTC m=+768.567346013" lastFinishedPulling="2025-11-23 06:58:11.269297117 +0000 UTC m=+793.291282730" observedRunningTime="2025-11-23 06:58:16.085489662 +0000 UTC m=+798.107475275" watchObservedRunningTime="2025-11-23 06:58:16.088973053 +0000 UTC m=+798.110958666" Nov 23 06:58:16 crc kubenswrapper[4559]: I1123 06:58:16.124672 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 23 06:58:16 crc kubenswrapper[4559]: I1123 06:58:16.124708 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 23 06:58:17 crc kubenswrapper[4559]: I1123 06:58:17.438565 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 23 06:58:17 crc kubenswrapper[4559]: I1123 06:58:17.438906 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 23 06:58:18 crc kubenswrapper[4559]: I1123 06:58:18.069248 4559 generic.go:334] "Generic (PLEG): container finished" podID="d534f227-972b-4a85-aeb5-8f1d226352e1" containerID="7bdd2d8e144c0d7b42f15b2a0273ab3c7b1897c9a15bc8d3ccab027508a84fe7" exitCode=0 Nov 23 06:58:18 crc kubenswrapper[4559]: I1123 06:58:18.069411 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jg7hk" event={"ID":"d534f227-972b-4a85-aeb5-8f1d226352e1","Type":"ContainerDied","Data":"7bdd2d8e144c0d7b42f15b2a0273ab3c7b1897c9a15bc8d3ccab027508a84fe7"} Nov 23 06:58:18 crc kubenswrapper[4559]: I1123 06:58:18.071808 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"5d19fe8b-239a-476a-9213-1eccbd58958e","Type":"ContainerStarted","Data":"e2349d8c35e5ca2042dcc96c1f5f417a8c14d20400966ae15bb2b35de9032825"} Nov 23 06:58:18 crc kubenswrapper[4559]: I1123 06:58:18.074079 4559 generic.go:334] "Generic (PLEG): container finished" podID="56c603bf-2cb8-4c83-b11f-7794dcc0114a" containerID="616b63b0514c438639d975b5dac16185db3e95d1a993bb586c1fd317be9f122e" exitCode=0 Nov 23 06:58:18 crc kubenswrapper[4559]: I1123 06:58:18.074172 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9v64g" event={"ID":"56c603bf-2cb8-4c83-b11f-7794dcc0114a","Type":"ContainerDied","Data":"616b63b0514c438639d975b5dac16185db3e95d1a993bb586c1fd317be9f122e"} Nov 23 06:58:18 crc kubenswrapper[4559]: I1123 06:58:18.075441 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f162538d-e942-4d5e-b2c5-11ba57054cf4","Type":"ContainerStarted","Data":"b895bfb12fa3225b7bfb1a792569a63e0d2d9a21078656b118dc7038e8e610df"} Nov 23 06:58:18 crc kubenswrapper[4559]: I1123 06:58:18.077863 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-blj85" event={"ID":"20662e83-ba79-4c97-80fb-98fda28c1149","Type":"ContainerStarted","Data":"23a163a08113c27b1aae8d511f3820c4323bff589d9f6fc61d559c3dcd1e9925"} Nov 23 06:58:18 crc kubenswrapper[4559]: I1123 06:58:18.119954 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-blj85" podStartSLOduration=19.477505515 podStartE2EDuration="25.119936373s" podCreationTimestamp="2025-11-23 06:57:53 +0000 UTC" firstStartedPulling="2025-11-23 06:58:11.7655326 +0000 UTC m=+793.787518213" lastFinishedPulling="2025-11-23 06:58:17.407963457 +0000 UTC m=+799.429949071" observedRunningTime="2025-11-23 06:58:18.118324552 +0000 UTC m=+800.140310166" watchObservedRunningTime="2025-11-23 06:58:18.119936373 +0000 UTC m=+800.141921988" Nov 23 06:58:18 crc kubenswrapper[4559]: I1123 06:58:18.176102 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7f47fdfb89-xjlhn" Nov 23 06:58:18 crc kubenswrapper[4559]: I1123 06:58:18.414169 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-blj85" Nov 23 06:58:18 crc kubenswrapper[4559]: I1123 06:58:18.460761 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-759c6cc4df-s47tx" Nov 23 06:58:18 crc kubenswrapper[4559]: I1123 06:58:18.511808 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f47fdfb89-xjlhn"] Nov 23 06:58:19 crc kubenswrapper[4559]: I1123 06:58:19.108410 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9v64g" event={"ID":"56c603bf-2cb8-4c83-b11f-7794dcc0114a","Type":"ContainerStarted","Data":"fc485fdda093cffab461870b19705161f65b8f27e38041fd0ded773bf0ebc799"} Nov 23 06:58:19 crc kubenswrapper[4559]: I1123 06:58:19.121218 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7f47fdfb89-xjlhn" podUID="ebba232a-99bb-4f45-a0ce-3628f42dcc77" containerName="dnsmasq-dns" containerID="cri-o://3e21393020565de7f1e1fe3808086932a925cb97580ccf93752fd78328a9da20" gracePeriod=10 Nov 23 06:58:19 crc kubenswrapper[4559]: I1123 06:58:19.123387 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jg7hk" event={"ID":"d534f227-972b-4a85-aeb5-8f1d226352e1","Type":"ContainerStarted","Data":"7ffa4325aef6eb60ff76ffbc53192feed59c52b40a06cd871f380ec9ccdc344f"} Nov 23 06:58:19 crc kubenswrapper[4559]: I1123 06:58:19.123417 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-jg7hk" Nov 23 06:58:19 crc kubenswrapper[4559]: I1123 06:58:19.123429 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jg7hk" event={"ID":"d534f227-972b-4a85-aeb5-8f1d226352e1","Type":"ContainerStarted","Data":"74e59139534f2ab1c8aadad1171bfdacb46da1ad1e70fb8391d5cea342c38b2e"} Nov 23 06:58:19 crc kubenswrapper[4559]: I1123 06:58:19.123967 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-jg7hk" Nov 23 06:58:19 crc kubenswrapper[4559]: I1123 06:58:19.128921 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9v64g" podStartSLOduration=3.457355232 podStartE2EDuration="5.1289046s" podCreationTimestamp="2025-11-23 06:58:14 +0000 UTC" firstStartedPulling="2025-11-23 06:58:16.91684253 +0000 UTC m=+798.938828154" lastFinishedPulling="2025-11-23 06:58:18.588391908 +0000 UTC m=+800.610377522" observedRunningTime="2025-11-23 06:58:19.123429896 +0000 UTC m=+801.145415509" watchObservedRunningTime="2025-11-23 06:58:19.1289046 +0000 UTC m=+801.150890215" Nov 23 06:58:19 crc kubenswrapper[4559]: I1123 06:58:19.148163 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-jg7hk" podStartSLOduration=20.614215459 podStartE2EDuration="26.148153495s" podCreationTimestamp="2025-11-23 06:57:53 +0000 UTC" firstStartedPulling="2025-11-23 06:58:11.85682181 +0000 UTC m=+793.878807424" lastFinishedPulling="2025-11-23 06:58:17.390759846 +0000 UTC m=+799.412745460" observedRunningTime="2025-11-23 06:58:19.142613629 +0000 UTC m=+801.164599243" watchObservedRunningTime="2025-11-23 06:58:19.148153495 +0000 UTC m=+801.170139110" Nov 23 06:58:19 crc kubenswrapper[4559]: I1123 06:58:19.734864 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 23 06:58:19 crc kubenswrapper[4559]: I1123 06:58:19.825477 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 23 06:58:20 crc kubenswrapper[4559]: I1123 06:58:20.131368 4559 generic.go:334] "Generic (PLEG): container finished" podID="ebba232a-99bb-4f45-a0ce-3628f42dcc77" containerID="3e21393020565de7f1e1fe3808086932a925cb97580ccf93752fd78328a9da20" exitCode=0 Nov 23 06:58:20 crc kubenswrapper[4559]: I1123 06:58:20.131411 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f47fdfb89-xjlhn" event={"ID":"ebba232a-99bb-4f45-a0ce-3628f42dcc77","Type":"ContainerDied","Data":"3e21393020565de7f1e1fe3808086932a925cb97580ccf93752fd78328a9da20"} Nov 23 06:58:20 crc kubenswrapper[4559]: I1123 06:58:20.622373 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f47fdfb89-xjlhn" Nov 23 06:58:20 crc kubenswrapper[4559]: I1123 06:58:20.780632 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ebba232a-99bb-4f45-a0ce-3628f42dcc77-dns-svc\") pod \"ebba232a-99bb-4f45-a0ce-3628f42dcc77\" (UID: \"ebba232a-99bb-4f45-a0ce-3628f42dcc77\") " Nov 23 06:58:20 crc kubenswrapper[4559]: I1123 06:58:20.780713 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m62j4\" (UniqueName: \"kubernetes.io/projected/ebba232a-99bb-4f45-a0ce-3628f42dcc77-kube-api-access-m62j4\") pod \"ebba232a-99bb-4f45-a0ce-3628f42dcc77\" (UID: \"ebba232a-99bb-4f45-a0ce-3628f42dcc77\") " Nov 23 06:58:20 crc kubenswrapper[4559]: I1123 06:58:20.780741 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebba232a-99bb-4f45-a0ce-3628f42dcc77-config\") pod \"ebba232a-99bb-4f45-a0ce-3628f42dcc77\" (UID: \"ebba232a-99bb-4f45-a0ce-3628f42dcc77\") " Nov 23 06:58:20 crc kubenswrapper[4559]: I1123 06:58:20.787140 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebba232a-99bb-4f45-a0ce-3628f42dcc77-kube-api-access-m62j4" (OuterVolumeSpecName: "kube-api-access-m62j4") pod "ebba232a-99bb-4f45-a0ce-3628f42dcc77" (UID: "ebba232a-99bb-4f45-a0ce-3628f42dcc77"). InnerVolumeSpecName "kube-api-access-m62j4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:58:20 crc kubenswrapper[4559]: I1123 06:58:20.813765 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebba232a-99bb-4f45-a0ce-3628f42dcc77-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ebba232a-99bb-4f45-a0ce-3628f42dcc77" (UID: "ebba232a-99bb-4f45-a0ce-3628f42dcc77"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:58:20 crc kubenswrapper[4559]: I1123 06:58:20.827980 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebba232a-99bb-4f45-a0ce-3628f42dcc77-config" (OuterVolumeSpecName: "config") pod "ebba232a-99bb-4f45-a0ce-3628f42dcc77" (UID: "ebba232a-99bb-4f45-a0ce-3628f42dcc77"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:58:20 crc kubenswrapper[4559]: I1123 06:58:20.884260 4559 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ebba232a-99bb-4f45-a0ce-3628f42dcc77-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:20 crc kubenswrapper[4559]: I1123 06:58:20.884291 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m62j4\" (UniqueName: \"kubernetes.io/projected/ebba232a-99bb-4f45-a0ce-3628f42dcc77-kube-api-access-m62j4\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:20 crc kubenswrapper[4559]: I1123 06:58:20.884304 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebba232a-99bb-4f45-a0ce-3628f42dcc77-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:21 crc kubenswrapper[4559]: I1123 06:58:21.140807 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f47fdfb89-xjlhn" event={"ID":"ebba232a-99bb-4f45-a0ce-3628f42dcc77","Type":"ContainerDied","Data":"10bed732f0acea79aff1443fa19dd7bdf6eeede05443ef73df76cd16a1dde70f"} Nov 23 06:58:21 crc kubenswrapper[4559]: I1123 06:58:21.140868 4559 scope.go:117] "RemoveContainer" containerID="3e21393020565de7f1e1fe3808086932a925cb97580ccf93752fd78328a9da20" Nov 23 06:58:21 crc kubenswrapper[4559]: I1123 06:58:21.141100 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f47fdfb89-xjlhn" Nov 23 06:58:21 crc kubenswrapper[4559]: I1123 06:58:21.143350 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"5d19fe8b-239a-476a-9213-1eccbd58958e","Type":"ContainerStarted","Data":"c7a1c2b955fd781043829be60e6a1aff544b053b38f29e1bf391b27cde083285"} Nov 23 06:58:21 crc kubenswrapper[4559]: I1123 06:58:21.147421 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f162538d-e942-4d5e-b2c5-11ba57054cf4","Type":"ContainerStarted","Data":"c8857a012e0098b62a5f5f05f7c0cc63c05e1731ce92f84a19e1fdc0b9dc96e4"} Nov 23 06:58:21 crc kubenswrapper[4559]: I1123 06:58:21.162023 4559 scope.go:117] "RemoveContainer" containerID="b21e2437402d5251e7379fd98e040e46e24f6b535fb2e9369a5171dfdc104d67" Nov 23 06:58:21 crc kubenswrapper[4559]: I1123 06:58:21.173526 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=20.119722579 podStartE2EDuration="29.173499834s" podCreationTimestamp="2025-11-23 06:57:52 +0000 UTC" firstStartedPulling="2025-11-23 06:58:11.764471194 +0000 UTC m=+793.786456808" lastFinishedPulling="2025-11-23 06:58:20.818248448 +0000 UTC m=+802.840234063" observedRunningTime="2025-11-23 06:58:21.163019888 +0000 UTC m=+803.185005502" watchObservedRunningTime="2025-11-23 06:58:21.173499834 +0000 UTC m=+803.195485448" Nov 23 06:58:21 crc kubenswrapper[4559]: I1123 06:58:21.180136 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f47fdfb89-xjlhn"] Nov 23 06:58:21 crc kubenswrapper[4559]: I1123 06:58:21.184229 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f47fdfb89-xjlhn"] Nov 23 06:58:21 crc kubenswrapper[4559]: I1123 06:58:21.195465 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=18.146149907 podStartE2EDuration="26.195433719s" podCreationTimestamp="2025-11-23 06:57:55 +0000 UTC" firstStartedPulling="2025-11-23 06:58:12.734631577 +0000 UTC m=+794.756617191" lastFinishedPulling="2025-11-23 06:58:20.783915388 +0000 UTC m=+802.805901003" observedRunningTime="2025-11-23 06:58:21.193253218 +0000 UTC m=+803.215238833" watchObservedRunningTime="2025-11-23 06:58:21.195433719 +0000 UTC m=+803.217419333" Nov 23 06:58:21 crc kubenswrapper[4559]: I1123 06:58:21.208509 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 23 06:58:21 crc kubenswrapper[4559]: I1123 06:58:21.238023 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 23 06:58:21 crc kubenswrapper[4559]: I1123 06:58:21.305408 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 23 06:58:21 crc kubenswrapper[4559]: I1123 06:58:21.332723 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.159242 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20","Type":"ContainerStarted","Data":"832a304e7da998cd94f31c322c8f84bedc01786f45ac24347d345e9f99aab1a6"} Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.162478 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.162574 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.194693 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.200291 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.206858 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.280441 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebba232a-99bb-4f45-a0ce-3628f42dcc77" path="/var/lib/kubelet/pods/ebba232a-99bb-4f45-a0ce-3628f42dcc77/volumes" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.281154 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.391057 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84d4c64565-8tdqj"] Nov 23 06:58:22 crc kubenswrapper[4559]: E1123 06:58:22.391352 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebba232a-99bb-4f45-a0ce-3628f42dcc77" containerName="dnsmasq-dns" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.391366 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebba232a-99bb-4f45-a0ce-3628f42dcc77" containerName="dnsmasq-dns" Nov 23 06:58:22 crc kubenswrapper[4559]: E1123 06:58:22.391393 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebba232a-99bb-4f45-a0ce-3628f42dcc77" containerName="init" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.391398 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebba232a-99bb-4f45-a0ce-3628f42dcc77" containerName="init" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.391548 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebba232a-99bb-4f45-a0ce-3628f42dcc77" containerName="dnsmasq-dns" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.392345 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84d4c64565-8tdqj" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.394757 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.402566 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84d4c64565-8tdqj"] Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.516868 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/767b5e8a-5df6-460e-9056-3ac035d16970-dns-svc\") pod \"dnsmasq-dns-84d4c64565-8tdqj\" (UID: \"767b5e8a-5df6-460e-9056-3ac035d16970\") " pod="openstack/dnsmasq-dns-84d4c64565-8tdqj" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.516944 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnkvc\" (UniqueName: \"kubernetes.io/projected/767b5e8a-5df6-460e-9056-3ac035d16970-kube-api-access-dnkvc\") pod \"dnsmasq-dns-84d4c64565-8tdqj\" (UID: \"767b5e8a-5df6-460e-9056-3ac035d16970\") " pod="openstack/dnsmasq-dns-84d4c64565-8tdqj" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.516998 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/767b5e8a-5df6-460e-9056-3ac035d16970-config\") pod \"dnsmasq-dns-84d4c64565-8tdqj\" (UID: \"767b5e8a-5df6-460e-9056-3ac035d16970\") " pod="openstack/dnsmasq-dns-84d4c64565-8tdqj" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.517224 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/767b5e8a-5df6-460e-9056-3ac035d16970-ovsdbserver-nb\") pod \"dnsmasq-dns-84d4c64565-8tdqj\" (UID: \"767b5e8a-5df6-460e-9056-3ac035d16970\") " pod="openstack/dnsmasq-dns-84d4c64565-8tdqj" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.572482 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-nzw6k"] Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.573570 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-nzw6k" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.583080 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-nzw6k"] Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.585993 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.618601 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnkvc\" (UniqueName: \"kubernetes.io/projected/767b5e8a-5df6-460e-9056-3ac035d16970-kube-api-access-dnkvc\") pod \"dnsmasq-dns-84d4c64565-8tdqj\" (UID: \"767b5e8a-5df6-460e-9056-3ac035d16970\") " pod="openstack/dnsmasq-dns-84d4c64565-8tdqj" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.618690 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/767b5e8a-5df6-460e-9056-3ac035d16970-config\") pod \"dnsmasq-dns-84d4c64565-8tdqj\" (UID: \"767b5e8a-5df6-460e-9056-3ac035d16970\") " pod="openstack/dnsmasq-dns-84d4c64565-8tdqj" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.618764 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/767b5e8a-5df6-460e-9056-3ac035d16970-ovsdbserver-nb\") pod \"dnsmasq-dns-84d4c64565-8tdqj\" (UID: \"767b5e8a-5df6-460e-9056-3ac035d16970\") " pod="openstack/dnsmasq-dns-84d4c64565-8tdqj" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.618951 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/767b5e8a-5df6-460e-9056-3ac035d16970-dns-svc\") pod \"dnsmasq-dns-84d4c64565-8tdqj\" (UID: \"767b5e8a-5df6-460e-9056-3ac035d16970\") " pod="openstack/dnsmasq-dns-84d4c64565-8tdqj" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.619696 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/767b5e8a-5df6-460e-9056-3ac035d16970-ovsdbserver-nb\") pod \"dnsmasq-dns-84d4c64565-8tdqj\" (UID: \"767b5e8a-5df6-460e-9056-3ac035d16970\") " pod="openstack/dnsmasq-dns-84d4c64565-8tdqj" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.619739 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/767b5e8a-5df6-460e-9056-3ac035d16970-config\") pod \"dnsmasq-dns-84d4c64565-8tdqj\" (UID: \"767b5e8a-5df6-460e-9056-3ac035d16970\") " pod="openstack/dnsmasq-dns-84d4c64565-8tdqj" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.619776 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/767b5e8a-5df6-460e-9056-3ac035d16970-dns-svc\") pod \"dnsmasq-dns-84d4c64565-8tdqj\" (UID: \"767b5e8a-5df6-460e-9056-3ac035d16970\") " pod="openstack/dnsmasq-dns-84d4c64565-8tdqj" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.634923 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84d4c64565-8tdqj"] Nov 23 06:58:22 crc kubenswrapper[4559]: E1123 06:58:22.635556 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-dnkvc], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-84d4c64565-8tdqj" podUID="767b5e8a-5df6-460e-9056-3ac035d16970" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.647714 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnkvc\" (UniqueName: \"kubernetes.io/projected/767b5e8a-5df6-460e-9056-3ac035d16970-kube-api-access-dnkvc\") pod \"dnsmasq-dns-84d4c64565-8tdqj\" (UID: \"767b5e8a-5df6-460e-9056-3ac035d16970\") " pod="openstack/dnsmasq-dns-84d4c64565-8tdqj" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.663202 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.664625 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.666405 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.666603 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.666700 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-7bbdr" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.667349 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.668448 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b8455895f-w72ch"] Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.669710 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b8455895f-w72ch" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.670973 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.676118 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.681707 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b8455895f-w72ch"] Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.720738 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-nzw6k\" (UID: \"dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5\") " pod="openstack/ovn-controller-metrics-nzw6k" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.720835 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5-config\") pod \"ovn-controller-metrics-nzw6k\" (UID: \"dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5\") " pod="openstack/ovn-controller-metrics-nzw6k" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.720892 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h2k2\" (UniqueName: \"kubernetes.io/projected/dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5-kube-api-access-9h2k2\") pod \"ovn-controller-metrics-nzw6k\" (UID: \"dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5\") " pod="openstack/ovn-controller-metrics-nzw6k" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.720994 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5-combined-ca-bundle\") pod \"ovn-controller-metrics-nzw6k\" (UID: \"dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5\") " pod="openstack/ovn-controller-metrics-nzw6k" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.721057 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5-ovn-rundir\") pod \"ovn-controller-metrics-nzw6k\" (UID: \"dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5\") " pod="openstack/ovn-controller-metrics-nzw6k" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.721091 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5-ovs-rundir\") pod \"ovn-controller-metrics-nzw6k\" (UID: \"dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5\") " pod="openstack/ovn-controller-metrics-nzw6k" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.822968 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjckk\" (UniqueName: \"kubernetes.io/projected/4460f104-a577-49e6-bb5a-dc50cb3401eb-kube-api-access-vjckk\") pod \"dnsmasq-dns-5b8455895f-w72ch\" (UID: \"4460f104-a577-49e6-bb5a-dc50cb3401eb\") " pod="openstack/dnsmasq-dns-5b8455895f-w72ch" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.823028 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4460f104-a577-49e6-bb5a-dc50cb3401eb-dns-svc\") pod \"dnsmasq-dns-5b8455895f-w72ch\" (UID: \"4460f104-a577-49e6-bb5a-dc50cb3401eb\") " pod="openstack/dnsmasq-dns-5b8455895f-w72ch" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.823097 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/00fa46d4-5d8e-43a6-a182-34faf0d694ab-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"00fa46d4-5d8e-43a6-a182-34faf0d694ab\") " pod="openstack/ovn-northd-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.823129 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-nzw6k\" (UID: \"dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5\") " pod="openstack/ovn-controller-metrics-nzw6k" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.823170 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbbsq\" (UniqueName: \"kubernetes.io/projected/00fa46d4-5d8e-43a6-a182-34faf0d694ab-kube-api-access-jbbsq\") pod \"ovn-northd-0\" (UID: \"00fa46d4-5d8e-43a6-a182-34faf0d694ab\") " pod="openstack/ovn-northd-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.823221 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5-config\") pod \"ovn-controller-metrics-nzw6k\" (UID: \"dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5\") " pod="openstack/ovn-controller-metrics-nzw6k" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.823253 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00fa46d4-5d8e-43a6-a182-34faf0d694ab-config\") pod \"ovn-northd-0\" (UID: \"00fa46d4-5d8e-43a6-a182-34faf0d694ab\") " pod="openstack/ovn-northd-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.823271 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4460f104-a577-49e6-bb5a-dc50cb3401eb-config\") pod \"dnsmasq-dns-5b8455895f-w72ch\" (UID: \"4460f104-a577-49e6-bb5a-dc50cb3401eb\") " pod="openstack/dnsmasq-dns-5b8455895f-w72ch" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.823292 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00fa46d4-5d8e-43a6-a182-34faf0d694ab-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"00fa46d4-5d8e-43a6-a182-34faf0d694ab\") " pod="openstack/ovn-northd-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.823313 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h2k2\" (UniqueName: \"kubernetes.io/projected/dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5-kube-api-access-9h2k2\") pod \"ovn-controller-metrics-nzw6k\" (UID: \"dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5\") " pod="openstack/ovn-controller-metrics-nzw6k" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.823331 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/00fa46d4-5d8e-43a6-a182-34faf0d694ab-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"00fa46d4-5d8e-43a6-a182-34faf0d694ab\") " pod="openstack/ovn-northd-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.823363 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/00fa46d4-5d8e-43a6-a182-34faf0d694ab-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"00fa46d4-5d8e-43a6-a182-34faf0d694ab\") " pod="openstack/ovn-northd-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.823408 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4460f104-a577-49e6-bb5a-dc50cb3401eb-ovsdbserver-sb\") pod \"dnsmasq-dns-5b8455895f-w72ch\" (UID: \"4460f104-a577-49e6-bb5a-dc50cb3401eb\") " pod="openstack/dnsmasq-dns-5b8455895f-w72ch" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.823433 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5-combined-ca-bundle\") pod \"ovn-controller-metrics-nzw6k\" (UID: \"dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5\") " pod="openstack/ovn-controller-metrics-nzw6k" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.823461 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00fa46d4-5d8e-43a6-a182-34faf0d694ab-scripts\") pod \"ovn-northd-0\" (UID: \"00fa46d4-5d8e-43a6-a182-34faf0d694ab\") " pod="openstack/ovn-northd-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.823479 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4460f104-a577-49e6-bb5a-dc50cb3401eb-ovsdbserver-nb\") pod \"dnsmasq-dns-5b8455895f-w72ch\" (UID: \"4460f104-a577-49e6-bb5a-dc50cb3401eb\") " pod="openstack/dnsmasq-dns-5b8455895f-w72ch" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.823524 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5-ovn-rundir\") pod \"ovn-controller-metrics-nzw6k\" (UID: \"dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5\") " pod="openstack/ovn-controller-metrics-nzw6k" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.823542 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5-ovs-rundir\") pod \"ovn-controller-metrics-nzw6k\" (UID: \"dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5\") " pod="openstack/ovn-controller-metrics-nzw6k" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.824008 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5-ovn-rundir\") pod \"ovn-controller-metrics-nzw6k\" (UID: \"dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5\") " pod="openstack/ovn-controller-metrics-nzw6k" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.824057 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5-ovs-rundir\") pod \"ovn-controller-metrics-nzw6k\" (UID: \"dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5\") " pod="openstack/ovn-controller-metrics-nzw6k" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.824568 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5-config\") pod \"ovn-controller-metrics-nzw6k\" (UID: \"dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5\") " pod="openstack/ovn-controller-metrics-nzw6k" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.828177 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-nzw6k\" (UID: \"dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5\") " pod="openstack/ovn-controller-metrics-nzw6k" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.843124 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5-combined-ca-bundle\") pod \"ovn-controller-metrics-nzw6k\" (UID: \"dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5\") " pod="openstack/ovn-controller-metrics-nzw6k" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.865124 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h2k2\" (UniqueName: \"kubernetes.io/projected/dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5-kube-api-access-9h2k2\") pod \"ovn-controller-metrics-nzw6k\" (UID: \"dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5\") " pod="openstack/ovn-controller-metrics-nzw6k" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.891309 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-nzw6k" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.924954 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4460f104-a577-49e6-bb5a-dc50cb3401eb-ovsdbserver-sb\") pod \"dnsmasq-dns-5b8455895f-w72ch\" (UID: \"4460f104-a577-49e6-bb5a-dc50cb3401eb\") " pod="openstack/dnsmasq-dns-5b8455895f-w72ch" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.925029 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00fa46d4-5d8e-43a6-a182-34faf0d694ab-scripts\") pod \"ovn-northd-0\" (UID: \"00fa46d4-5d8e-43a6-a182-34faf0d694ab\") " pod="openstack/ovn-northd-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.925053 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4460f104-a577-49e6-bb5a-dc50cb3401eb-ovsdbserver-nb\") pod \"dnsmasq-dns-5b8455895f-w72ch\" (UID: \"4460f104-a577-49e6-bb5a-dc50cb3401eb\") " pod="openstack/dnsmasq-dns-5b8455895f-w72ch" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.925123 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjckk\" (UniqueName: \"kubernetes.io/projected/4460f104-a577-49e6-bb5a-dc50cb3401eb-kube-api-access-vjckk\") pod \"dnsmasq-dns-5b8455895f-w72ch\" (UID: \"4460f104-a577-49e6-bb5a-dc50cb3401eb\") " pod="openstack/dnsmasq-dns-5b8455895f-w72ch" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.925153 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4460f104-a577-49e6-bb5a-dc50cb3401eb-dns-svc\") pod \"dnsmasq-dns-5b8455895f-w72ch\" (UID: \"4460f104-a577-49e6-bb5a-dc50cb3401eb\") " pod="openstack/dnsmasq-dns-5b8455895f-w72ch" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.925207 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/00fa46d4-5d8e-43a6-a182-34faf0d694ab-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"00fa46d4-5d8e-43a6-a182-34faf0d694ab\") " pod="openstack/ovn-northd-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.925244 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbbsq\" (UniqueName: \"kubernetes.io/projected/00fa46d4-5d8e-43a6-a182-34faf0d694ab-kube-api-access-jbbsq\") pod \"ovn-northd-0\" (UID: \"00fa46d4-5d8e-43a6-a182-34faf0d694ab\") " pod="openstack/ovn-northd-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.925305 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00fa46d4-5d8e-43a6-a182-34faf0d694ab-config\") pod \"ovn-northd-0\" (UID: \"00fa46d4-5d8e-43a6-a182-34faf0d694ab\") " pod="openstack/ovn-northd-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.925329 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4460f104-a577-49e6-bb5a-dc50cb3401eb-config\") pod \"dnsmasq-dns-5b8455895f-w72ch\" (UID: \"4460f104-a577-49e6-bb5a-dc50cb3401eb\") " pod="openstack/dnsmasq-dns-5b8455895f-w72ch" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.925345 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00fa46d4-5d8e-43a6-a182-34faf0d694ab-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"00fa46d4-5d8e-43a6-a182-34faf0d694ab\") " pod="openstack/ovn-northd-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.925367 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/00fa46d4-5d8e-43a6-a182-34faf0d694ab-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"00fa46d4-5d8e-43a6-a182-34faf0d694ab\") " pod="openstack/ovn-northd-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.925396 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/00fa46d4-5d8e-43a6-a182-34faf0d694ab-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"00fa46d4-5d8e-43a6-a182-34faf0d694ab\") " pod="openstack/ovn-northd-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.925911 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4460f104-a577-49e6-bb5a-dc50cb3401eb-ovsdbserver-sb\") pod \"dnsmasq-dns-5b8455895f-w72ch\" (UID: \"4460f104-a577-49e6-bb5a-dc50cb3401eb\") " pod="openstack/dnsmasq-dns-5b8455895f-w72ch" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.925976 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00fa46d4-5d8e-43a6-a182-34faf0d694ab-scripts\") pod \"ovn-northd-0\" (UID: \"00fa46d4-5d8e-43a6-a182-34faf0d694ab\") " pod="openstack/ovn-northd-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.926087 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4460f104-a577-49e6-bb5a-dc50cb3401eb-ovsdbserver-nb\") pod \"dnsmasq-dns-5b8455895f-w72ch\" (UID: \"4460f104-a577-49e6-bb5a-dc50cb3401eb\") " pod="openstack/dnsmasq-dns-5b8455895f-w72ch" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.926208 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/00fa46d4-5d8e-43a6-a182-34faf0d694ab-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"00fa46d4-5d8e-43a6-a182-34faf0d694ab\") " pod="openstack/ovn-northd-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.926498 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4460f104-a577-49e6-bb5a-dc50cb3401eb-config\") pod \"dnsmasq-dns-5b8455895f-w72ch\" (UID: \"4460f104-a577-49e6-bb5a-dc50cb3401eb\") " pod="openstack/dnsmasq-dns-5b8455895f-w72ch" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.926851 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00fa46d4-5d8e-43a6-a182-34faf0d694ab-config\") pod \"ovn-northd-0\" (UID: \"00fa46d4-5d8e-43a6-a182-34faf0d694ab\") " pod="openstack/ovn-northd-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.927524 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4460f104-a577-49e6-bb5a-dc50cb3401eb-dns-svc\") pod \"dnsmasq-dns-5b8455895f-w72ch\" (UID: \"4460f104-a577-49e6-bb5a-dc50cb3401eb\") " pod="openstack/dnsmasq-dns-5b8455895f-w72ch" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.929166 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/00fa46d4-5d8e-43a6-a182-34faf0d694ab-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"00fa46d4-5d8e-43a6-a182-34faf0d694ab\") " pod="openstack/ovn-northd-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.933074 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00fa46d4-5d8e-43a6-a182-34faf0d694ab-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"00fa46d4-5d8e-43a6-a182-34faf0d694ab\") " pod="openstack/ovn-northd-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.933675 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/00fa46d4-5d8e-43a6-a182-34faf0d694ab-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"00fa46d4-5d8e-43a6-a182-34faf0d694ab\") " pod="openstack/ovn-northd-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.939097 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjckk\" (UniqueName: \"kubernetes.io/projected/4460f104-a577-49e6-bb5a-dc50cb3401eb-kube-api-access-vjckk\") pod \"dnsmasq-dns-5b8455895f-w72ch\" (UID: \"4460f104-a577-49e6-bb5a-dc50cb3401eb\") " pod="openstack/dnsmasq-dns-5b8455895f-w72ch" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.941384 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbbsq\" (UniqueName: \"kubernetes.io/projected/00fa46d4-5d8e-43a6-a182-34faf0d694ab-kube-api-access-jbbsq\") pod \"ovn-northd-0\" (UID: \"00fa46d4-5d8e-43a6-a182-34faf0d694ab\") " pod="openstack/ovn-northd-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.981354 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 23 06:58:22 crc kubenswrapper[4559]: I1123 06:58:22.988572 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b8455895f-w72ch" Nov 23 06:58:23 crc kubenswrapper[4559]: I1123 06:58:23.168552 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84d4c64565-8tdqj" Nov 23 06:58:23 crc kubenswrapper[4559]: I1123 06:58:23.177457 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84d4c64565-8tdqj" Nov 23 06:58:23 crc kubenswrapper[4559]: I1123 06:58:23.288120 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-nzw6k"] Nov 23 06:58:23 crc kubenswrapper[4559]: I1123 06:58:23.331702 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnkvc\" (UniqueName: \"kubernetes.io/projected/767b5e8a-5df6-460e-9056-3ac035d16970-kube-api-access-dnkvc\") pod \"767b5e8a-5df6-460e-9056-3ac035d16970\" (UID: \"767b5e8a-5df6-460e-9056-3ac035d16970\") " Nov 23 06:58:23 crc kubenswrapper[4559]: I1123 06:58:23.331774 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/767b5e8a-5df6-460e-9056-3ac035d16970-ovsdbserver-nb\") pod \"767b5e8a-5df6-460e-9056-3ac035d16970\" (UID: \"767b5e8a-5df6-460e-9056-3ac035d16970\") " Nov 23 06:58:23 crc kubenswrapper[4559]: I1123 06:58:23.331927 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/767b5e8a-5df6-460e-9056-3ac035d16970-config\") pod \"767b5e8a-5df6-460e-9056-3ac035d16970\" (UID: \"767b5e8a-5df6-460e-9056-3ac035d16970\") " Nov 23 06:58:23 crc kubenswrapper[4559]: I1123 06:58:23.331972 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/767b5e8a-5df6-460e-9056-3ac035d16970-dns-svc\") pod \"767b5e8a-5df6-460e-9056-3ac035d16970\" (UID: \"767b5e8a-5df6-460e-9056-3ac035d16970\") " Nov 23 06:58:23 crc kubenswrapper[4559]: I1123 06:58:23.333900 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/767b5e8a-5df6-460e-9056-3ac035d16970-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "767b5e8a-5df6-460e-9056-3ac035d16970" (UID: "767b5e8a-5df6-460e-9056-3ac035d16970"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:58:23 crc kubenswrapper[4559]: I1123 06:58:23.334317 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/767b5e8a-5df6-460e-9056-3ac035d16970-config" (OuterVolumeSpecName: "config") pod "767b5e8a-5df6-460e-9056-3ac035d16970" (UID: "767b5e8a-5df6-460e-9056-3ac035d16970"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:58:23 crc kubenswrapper[4559]: I1123 06:58:23.334871 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/767b5e8a-5df6-460e-9056-3ac035d16970-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "767b5e8a-5df6-460e-9056-3ac035d16970" (UID: "767b5e8a-5df6-460e-9056-3ac035d16970"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:58:23 crc kubenswrapper[4559]: I1123 06:58:23.337269 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/767b5e8a-5df6-460e-9056-3ac035d16970-kube-api-access-dnkvc" (OuterVolumeSpecName: "kube-api-access-dnkvc") pod "767b5e8a-5df6-460e-9056-3ac035d16970" (UID: "767b5e8a-5df6-460e-9056-3ac035d16970"). InnerVolumeSpecName "kube-api-access-dnkvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:58:23 crc kubenswrapper[4559]: I1123 06:58:23.413326 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b8455895f-w72ch"] Nov 23 06:58:23 crc kubenswrapper[4559]: W1123 06:58:23.420436 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4460f104_a577_49e6_bb5a_dc50cb3401eb.slice/crio-e8d3d48cc45ee9c94b55291517a0d9e1e1e37f3f67309e35f84fd45f8abf570f WatchSource:0}: Error finding container e8d3d48cc45ee9c94b55291517a0d9e1e1e37f3f67309e35f84fd45f8abf570f: Status 404 returned error can't find the container with id e8d3d48cc45ee9c94b55291517a0d9e1e1e37f3f67309e35f84fd45f8abf570f Nov 23 06:58:23 crc kubenswrapper[4559]: I1123 06:58:23.434584 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnkvc\" (UniqueName: \"kubernetes.io/projected/767b5e8a-5df6-460e-9056-3ac035d16970-kube-api-access-dnkvc\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:23 crc kubenswrapper[4559]: I1123 06:58:23.434614 4559 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/767b5e8a-5df6-460e-9056-3ac035d16970-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:23 crc kubenswrapper[4559]: I1123 06:58:23.434627 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/767b5e8a-5df6-460e-9056-3ac035d16970-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:23 crc kubenswrapper[4559]: I1123 06:58:23.434636 4559 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/767b5e8a-5df6-460e-9056-3ac035d16970-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:23 crc kubenswrapper[4559]: I1123 06:58:23.460413 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 23 06:58:23 crc kubenswrapper[4559]: W1123 06:58:23.464002 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00fa46d4_5d8e_43a6_a182_34faf0d694ab.slice/crio-be7a10eb7d21962ca92d448fe92c505c331def5d838bed4742483a2565ce4eb4 WatchSource:0}: Error finding container be7a10eb7d21962ca92d448fe92c505c331def5d838bed4742483a2565ce4eb4: Status 404 returned error can't find the container with id be7a10eb7d21962ca92d448fe92c505c331def5d838bed4742483a2565ce4eb4 Nov 23 06:58:24 crc kubenswrapper[4559]: I1123 06:58:24.180592 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"00fa46d4-5d8e-43a6-a182-34faf0d694ab","Type":"ContainerStarted","Data":"be7a10eb7d21962ca92d448fe92c505c331def5d838bed4742483a2565ce4eb4"} Nov 23 06:58:24 crc kubenswrapper[4559]: I1123 06:58:24.183625 4559 generic.go:334] "Generic (PLEG): container finished" podID="4460f104-a577-49e6-bb5a-dc50cb3401eb" containerID="d656a10bef82c04d5729a7b40002f50b4178794e85a139ecb2cbf8435b833736" exitCode=0 Nov 23 06:58:24 crc kubenswrapper[4559]: I1123 06:58:24.183698 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b8455895f-w72ch" event={"ID":"4460f104-a577-49e6-bb5a-dc50cb3401eb","Type":"ContainerDied","Data":"d656a10bef82c04d5729a7b40002f50b4178794e85a139ecb2cbf8435b833736"} Nov 23 06:58:24 crc kubenswrapper[4559]: I1123 06:58:24.183717 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b8455895f-w72ch" event={"ID":"4460f104-a577-49e6-bb5a-dc50cb3401eb","Type":"ContainerStarted","Data":"e8d3d48cc45ee9c94b55291517a0d9e1e1e37f3f67309e35f84fd45f8abf570f"} Nov 23 06:58:24 crc kubenswrapper[4559]: I1123 06:58:24.187926 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-nzw6k" event={"ID":"dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5","Type":"ContainerStarted","Data":"1c0390cacba0de1765162d192c85bf99599341a916a4c058e377d9ef218fbbae"} Nov 23 06:58:24 crc kubenswrapper[4559]: I1123 06:58:24.187967 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-nzw6k" event={"ID":"dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5","Type":"ContainerStarted","Data":"ebdac0e3520bdb46210f3970903d050ef42ccf54e76ea738938f9615445be750"} Nov 23 06:58:24 crc kubenswrapper[4559]: I1123 06:58:24.188484 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84d4c64565-8tdqj" Nov 23 06:58:24 crc kubenswrapper[4559]: I1123 06:58:24.230707 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-nzw6k" podStartSLOduration=2.230688503 podStartE2EDuration="2.230688503s" podCreationTimestamp="2025-11-23 06:58:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:58:24.223840777 +0000 UTC m=+806.245826391" watchObservedRunningTime="2025-11-23 06:58:24.230688503 +0000 UTC m=+806.252674117" Nov 23 06:58:24 crc kubenswrapper[4559]: I1123 06:58:24.325343 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84d4c64565-8tdqj"] Nov 23 06:58:24 crc kubenswrapper[4559]: I1123 06:58:24.332804 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84d4c64565-8tdqj"] Nov 23 06:58:25 crc kubenswrapper[4559]: I1123 06:58:25.068724 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9v64g" Nov 23 06:58:25 crc kubenswrapper[4559]: I1123 06:58:25.068774 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9v64g" Nov 23 06:58:25 crc kubenswrapper[4559]: I1123 06:58:25.102260 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9v64g" Nov 23 06:58:25 crc kubenswrapper[4559]: I1123 06:58:25.263186 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9v64g" Nov 23 06:58:25 crc kubenswrapper[4559]: I1123 06:58:25.324982 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9v64g"] Nov 23 06:58:26 crc kubenswrapper[4559]: I1123 06:58:26.215615 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b8455895f-w72ch" event={"ID":"4460f104-a577-49e6-bb5a-dc50cb3401eb","Type":"ContainerStarted","Data":"a0141bb4b5a5004234d5f211f1fb42f3f155c25253aa7cce564a0e938e8bc619"} Nov 23 06:58:26 crc kubenswrapper[4559]: I1123 06:58:26.215784 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b8455895f-w72ch" Nov 23 06:58:26 crc kubenswrapper[4559]: I1123 06:58:26.218848 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"de0cb0c4-ffde-44c5-adb9-7aea0692c1b5","Type":"ContainerStarted","Data":"7e3c890de64947899b0ec310f4b37145c710dd9831451fc8ed6371bf24d33ef8"} Nov 23 06:58:26 crc kubenswrapper[4559]: I1123 06:58:26.219259 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 23 06:58:26 crc kubenswrapper[4559]: I1123 06:58:26.222258 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"61e2f87a-db68-432a-8d13-94087d1afb17","Type":"ContainerStarted","Data":"3ed96cde7ed68fbff6a9ee5c583b3e101eecb01f52f5e3d5f2ca7b0c38bf7687"} Nov 23 06:58:26 crc kubenswrapper[4559]: I1123 06:58:26.236138 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b8455895f-w72ch" podStartSLOduration=4.236120764 podStartE2EDuration="4.236120764s" podCreationTimestamp="2025-11-23 06:58:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:58:26.229661319 +0000 UTC m=+808.251646922" watchObservedRunningTime="2025-11-23 06:58:26.236120764 +0000 UTC m=+808.258106379" Nov 23 06:58:26 crc kubenswrapper[4559]: I1123 06:58:26.263170 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=1.8361807049999999 podStartE2EDuration="39.263152135s" podCreationTimestamp="2025-11-23 06:57:47 +0000 UTC" firstStartedPulling="2025-11-23 06:57:48.42884549 +0000 UTC m=+770.450831094" lastFinishedPulling="2025-11-23 06:58:25.85581691 +0000 UTC m=+807.877802524" observedRunningTime="2025-11-23 06:58:26.259550843 +0000 UTC m=+808.281536457" watchObservedRunningTime="2025-11-23 06:58:26.263152135 +0000 UTC m=+808.285137749" Nov 23 06:58:26 crc kubenswrapper[4559]: I1123 06:58:26.282740 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="767b5e8a-5df6-460e-9056-3ac035d16970" path="/var/lib/kubelet/pods/767b5e8a-5df6-460e-9056-3ac035d16970/volumes" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.229047 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"00fa46d4-5d8e-43a6-a182-34faf0d694ab","Type":"ContainerStarted","Data":"cb5472c0542e008486962fb3d30d223ac6a02ccd55eded332735fc18cccefb61"} Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.229241 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"00fa46d4-5d8e-43a6-a182-34faf0d694ab","Type":"ContainerStarted","Data":"e1dfeb186b03526aeea705b3f27383e9b30ea867bbdcbe8967693f1cbe811a0f"} Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.229224 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9v64g" podUID="56c603bf-2cb8-4c83-b11f-7794dcc0114a" containerName="registry-server" containerID="cri-o://fc485fdda093cffab461870b19705161f65b8f27e38041fd0ded773bf0ebc799" gracePeriod=2 Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.253375 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.407446057 podStartE2EDuration="5.253352871s" podCreationTimestamp="2025-11-23 06:58:22 +0000 UTC" firstStartedPulling="2025-11-23 06:58:23.466047078 +0000 UTC m=+805.488032692" lastFinishedPulling="2025-11-23 06:58:26.311953891 +0000 UTC m=+808.333939506" observedRunningTime="2025-11-23 06:58:27.249414556 +0000 UTC m=+809.271400170" watchObservedRunningTime="2025-11-23 06:58:27.253352871 +0000 UTC m=+809.275338485" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.593060 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9v64g" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.614372 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-c1ff-account-create-ngpwh"] Nov 23 06:58:27 crc kubenswrapper[4559]: E1123 06:58:27.614726 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56c603bf-2cb8-4c83-b11f-7794dcc0114a" containerName="extract-utilities" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.614743 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="56c603bf-2cb8-4c83-b11f-7794dcc0114a" containerName="extract-utilities" Nov 23 06:58:27 crc kubenswrapper[4559]: E1123 06:58:27.614760 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56c603bf-2cb8-4c83-b11f-7794dcc0114a" containerName="extract-content" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.614767 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="56c603bf-2cb8-4c83-b11f-7794dcc0114a" containerName="extract-content" Nov 23 06:58:27 crc kubenswrapper[4559]: E1123 06:58:27.614791 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56c603bf-2cb8-4c83-b11f-7794dcc0114a" containerName="registry-server" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.614797 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="56c603bf-2cb8-4c83-b11f-7794dcc0114a" containerName="registry-server" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.614956 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="56c603bf-2cb8-4c83-b11f-7794dcc0114a" containerName="registry-server" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.615448 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c1ff-account-create-ngpwh" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.617567 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.630956 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-c1ff-account-create-ngpwh"] Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.641242 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-48ksj"] Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.642196 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-48ksj" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.655612 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-48ksj"] Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.706000 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56c603bf-2cb8-4c83-b11f-7794dcc0114a-catalog-content\") pod \"56c603bf-2cb8-4c83-b11f-7794dcc0114a\" (UID: \"56c603bf-2cb8-4c83-b11f-7794dcc0114a\") " Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.706063 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqm4q\" (UniqueName: \"kubernetes.io/projected/56c603bf-2cb8-4c83-b11f-7794dcc0114a-kube-api-access-rqm4q\") pod \"56c603bf-2cb8-4c83-b11f-7794dcc0114a\" (UID: \"56c603bf-2cb8-4c83-b11f-7794dcc0114a\") " Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.706177 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56c603bf-2cb8-4c83-b11f-7794dcc0114a-utilities\") pod \"56c603bf-2cb8-4c83-b11f-7794dcc0114a\" (UID: \"56c603bf-2cb8-4c83-b11f-7794dcc0114a\") " Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.706458 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a81225a3-0928-42f9-bdc8-22375a649775-operator-scripts\") pod \"keystone-c1ff-account-create-ngpwh\" (UID: \"a81225a3-0928-42f9-bdc8-22375a649775\") " pod="openstack/keystone-c1ff-account-create-ngpwh" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.706569 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de1904f5-a9a6-4bc2-b106-71cbc351650b-operator-scripts\") pod \"keystone-db-create-48ksj\" (UID: \"de1904f5-a9a6-4bc2-b106-71cbc351650b\") " pod="openstack/keystone-db-create-48ksj" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.706786 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkcvg\" (UniqueName: \"kubernetes.io/projected/a81225a3-0928-42f9-bdc8-22375a649775-kube-api-access-mkcvg\") pod \"keystone-c1ff-account-create-ngpwh\" (UID: \"a81225a3-0928-42f9-bdc8-22375a649775\") " pod="openstack/keystone-c1ff-account-create-ngpwh" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.706972 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdhx9\" (UniqueName: \"kubernetes.io/projected/de1904f5-a9a6-4bc2-b106-71cbc351650b-kube-api-access-wdhx9\") pod \"keystone-db-create-48ksj\" (UID: \"de1904f5-a9a6-4bc2-b106-71cbc351650b\") " pod="openstack/keystone-db-create-48ksj" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.707019 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56c603bf-2cb8-4c83-b11f-7794dcc0114a-utilities" (OuterVolumeSpecName: "utilities") pod "56c603bf-2cb8-4c83-b11f-7794dcc0114a" (UID: "56c603bf-2cb8-4c83-b11f-7794dcc0114a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.707257 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56c603bf-2cb8-4c83-b11f-7794dcc0114a-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.710297 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56c603bf-2cb8-4c83-b11f-7794dcc0114a-kube-api-access-rqm4q" (OuterVolumeSpecName: "kube-api-access-rqm4q") pod "56c603bf-2cb8-4c83-b11f-7794dcc0114a" (UID: "56c603bf-2cb8-4c83-b11f-7794dcc0114a"). InnerVolumeSpecName "kube-api-access-rqm4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.751385 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56c603bf-2cb8-4c83-b11f-7794dcc0114a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "56c603bf-2cb8-4c83-b11f-7794dcc0114a" (UID: "56c603bf-2cb8-4c83-b11f-7794dcc0114a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.808700 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdhx9\" (UniqueName: \"kubernetes.io/projected/de1904f5-a9a6-4bc2-b106-71cbc351650b-kube-api-access-wdhx9\") pod \"keystone-db-create-48ksj\" (UID: \"de1904f5-a9a6-4bc2-b106-71cbc351650b\") " pod="openstack/keystone-db-create-48ksj" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.808817 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a81225a3-0928-42f9-bdc8-22375a649775-operator-scripts\") pod \"keystone-c1ff-account-create-ngpwh\" (UID: \"a81225a3-0928-42f9-bdc8-22375a649775\") " pod="openstack/keystone-c1ff-account-create-ngpwh" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.808920 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de1904f5-a9a6-4bc2-b106-71cbc351650b-operator-scripts\") pod \"keystone-db-create-48ksj\" (UID: \"de1904f5-a9a6-4bc2-b106-71cbc351650b\") " pod="openstack/keystone-db-create-48ksj" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.808990 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkcvg\" (UniqueName: \"kubernetes.io/projected/a81225a3-0928-42f9-bdc8-22375a649775-kube-api-access-mkcvg\") pod \"keystone-c1ff-account-create-ngpwh\" (UID: \"a81225a3-0928-42f9-bdc8-22375a649775\") " pod="openstack/keystone-c1ff-account-create-ngpwh" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.809079 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56c603bf-2cb8-4c83-b11f-7794dcc0114a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.809096 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqm4q\" (UniqueName: \"kubernetes.io/projected/56c603bf-2cb8-4c83-b11f-7794dcc0114a-kube-api-access-rqm4q\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.809612 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a81225a3-0928-42f9-bdc8-22375a649775-operator-scripts\") pod \"keystone-c1ff-account-create-ngpwh\" (UID: \"a81225a3-0928-42f9-bdc8-22375a649775\") " pod="openstack/keystone-c1ff-account-create-ngpwh" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.810206 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de1904f5-a9a6-4bc2-b106-71cbc351650b-operator-scripts\") pod \"keystone-db-create-48ksj\" (UID: \"de1904f5-a9a6-4bc2-b106-71cbc351650b\") " pod="openstack/keystone-db-create-48ksj" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.823110 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkcvg\" (UniqueName: \"kubernetes.io/projected/a81225a3-0928-42f9-bdc8-22375a649775-kube-api-access-mkcvg\") pod \"keystone-c1ff-account-create-ngpwh\" (UID: \"a81225a3-0928-42f9-bdc8-22375a649775\") " pod="openstack/keystone-c1ff-account-create-ngpwh" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.823782 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdhx9\" (UniqueName: \"kubernetes.io/projected/de1904f5-a9a6-4bc2-b106-71cbc351650b-kube-api-access-wdhx9\") pod \"keystone-db-create-48ksj\" (UID: \"de1904f5-a9a6-4bc2-b106-71cbc351650b\") " pod="openstack/keystone-db-create-48ksj" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.837600 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-556xz"] Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.838784 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-556xz" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.846296 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-556xz"] Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.910303 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q52n2\" (UniqueName: \"kubernetes.io/projected/f26678a1-0d1d-40c6-9eef-15f8ff08caab-kube-api-access-q52n2\") pod \"placement-db-create-556xz\" (UID: \"f26678a1-0d1d-40c6-9eef-15f8ff08caab\") " pod="openstack/placement-db-create-556xz" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.910406 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f26678a1-0d1d-40c6-9eef-15f8ff08caab-operator-scripts\") pod \"placement-db-create-556xz\" (UID: \"f26678a1-0d1d-40c6-9eef-15f8ff08caab\") " pod="openstack/placement-db-create-556xz" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.933559 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c1ff-account-create-ngpwh" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.942282 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-9b47-account-create-br9bd"] Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.943284 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9b47-account-create-br9bd" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.944833 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.950751 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9b47-account-create-br9bd"] Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.958775 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-48ksj" Nov 23 06:58:27 crc kubenswrapper[4559]: I1123 06:58:27.982738 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.012060 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f26678a1-0d1d-40c6-9eef-15f8ff08caab-operator-scripts\") pod \"placement-db-create-556xz\" (UID: \"f26678a1-0d1d-40c6-9eef-15f8ff08caab\") " pod="openstack/placement-db-create-556xz" Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.012289 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q52n2\" (UniqueName: \"kubernetes.io/projected/f26678a1-0d1d-40c6-9eef-15f8ff08caab-kube-api-access-q52n2\") pod \"placement-db-create-556xz\" (UID: \"f26678a1-0d1d-40c6-9eef-15f8ff08caab\") " pod="openstack/placement-db-create-556xz" Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.012377 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2efe0b7a-32c9-40b6-a933-014c9561b25e-operator-scripts\") pod \"placement-9b47-account-create-br9bd\" (UID: \"2efe0b7a-32c9-40b6-a933-014c9561b25e\") " pod="openstack/placement-9b47-account-create-br9bd" Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.012495 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmcln\" (UniqueName: \"kubernetes.io/projected/2efe0b7a-32c9-40b6-a933-014c9561b25e-kube-api-access-qmcln\") pod \"placement-9b47-account-create-br9bd\" (UID: \"2efe0b7a-32c9-40b6-a933-014c9561b25e\") " pod="openstack/placement-9b47-account-create-br9bd" Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.013193 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f26678a1-0d1d-40c6-9eef-15f8ff08caab-operator-scripts\") pod \"placement-db-create-556xz\" (UID: \"f26678a1-0d1d-40c6-9eef-15f8ff08caab\") " pod="openstack/placement-db-create-556xz" Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.030226 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q52n2\" (UniqueName: \"kubernetes.io/projected/f26678a1-0d1d-40c6-9eef-15f8ff08caab-kube-api-access-q52n2\") pod \"placement-db-create-556xz\" (UID: \"f26678a1-0d1d-40c6-9eef-15f8ff08caab\") " pod="openstack/placement-db-create-556xz" Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.113448 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmcln\" (UniqueName: \"kubernetes.io/projected/2efe0b7a-32c9-40b6-a933-014c9561b25e-kube-api-access-qmcln\") pod \"placement-9b47-account-create-br9bd\" (UID: \"2efe0b7a-32c9-40b6-a933-014c9561b25e\") " pod="openstack/placement-9b47-account-create-br9bd" Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.113585 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2efe0b7a-32c9-40b6-a933-014c9561b25e-operator-scripts\") pod \"placement-9b47-account-create-br9bd\" (UID: \"2efe0b7a-32c9-40b6-a933-014c9561b25e\") " pod="openstack/placement-9b47-account-create-br9bd" Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.114160 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2efe0b7a-32c9-40b6-a933-014c9561b25e-operator-scripts\") pod \"placement-9b47-account-create-br9bd\" (UID: \"2efe0b7a-32c9-40b6-a933-014c9561b25e\") " pod="openstack/placement-9b47-account-create-br9bd" Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.141698 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmcln\" (UniqueName: \"kubernetes.io/projected/2efe0b7a-32c9-40b6-a933-014c9561b25e-kube-api-access-qmcln\") pod \"placement-9b47-account-create-br9bd\" (UID: \"2efe0b7a-32c9-40b6-a933-014c9561b25e\") " pod="openstack/placement-9b47-account-create-br9bd" Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.160066 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-556xz" Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.173538 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-48ksj"] Nov 23 06:58:28 crc kubenswrapper[4559]: W1123 06:58:28.177335 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde1904f5_a9a6_4bc2_b106_71cbc351650b.slice/crio-f9e73b0ab841045907311e0cefaa212dbc9e278213cee9bd31516be4bf282c91 WatchSource:0}: Error finding container f9e73b0ab841045907311e0cefaa212dbc9e278213cee9bd31516be4bf282c91: Status 404 returned error can't find the container with id f9e73b0ab841045907311e0cefaa212dbc9e278213cee9bd31516be4bf282c91 Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.247547 4559 generic.go:334] "Generic (PLEG): container finished" podID="56c603bf-2cb8-4c83-b11f-7794dcc0114a" containerID="fc485fdda093cffab461870b19705161f65b8f27e38041fd0ded773bf0ebc799" exitCode=0 Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.247615 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9v64g" event={"ID":"56c603bf-2cb8-4c83-b11f-7794dcc0114a","Type":"ContainerDied","Data":"fc485fdda093cffab461870b19705161f65b8f27e38041fd0ded773bf0ebc799"} Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.247661 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9v64g" event={"ID":"56c603bf-2cb8-4c83-b11f-7794dcc0114a","Type":"ContainerDied","Data":"a80fe4b5bbc3a21c49d0738d4bc88c06988a148d17f4fe09d7b2bf5d1d9b6d13"} Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.247683 4559 scope.go:117] "RemoveContainer" containerID="fc485fdda093cffab461870b19705161f65b8f27e38041fd0ded773bf0ebc799" Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.247839 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9v64g" Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.254032 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-48ksj" event={"ID":"de1904f5-a9a6-4bc2-b106-71cbc351650b","Type":"ContainerStarted","Data":"f9e73b0ab841045907311e0cefaa212dbc9e278213cee9bd31516be4bf282c91"} Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.279754 4559 scope.go:117] "RemoveContainer" containerID="616b63b0514c438639d975b5dac16185db3e95d1a993bb586c1fd317be9f122e" Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.308310 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9v64g"] Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.308357 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9v64g"] Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.309654 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9b47-account-create-br9bd" Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.311172 4559 scope.go:117] "RemoveContainer" containerID="481c1c0cdb7c672d4a4a53468a8b1360b3efaffa13d3dc3b3b8a200072b6173d" Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.343736 4559 scope.go:117] "RemoveContainer" containerID="fc485fdda093cffab461870b19705161f65b8f27e38041fd0ded773bf0ebc799" Nov 23 06:58:28 crc kubenswrapper[4559]: E1123 06:58:28.344708 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc485fdda093cffab461870b19705161f65b8f27e38041fd0ded773bf0ebc799\": container with ID starting with fc485fdda093cffab461870b19705161f65b8f27e38041fd0ded773bf0ebc799 not found: ID does not exist" containerID="fc485fdda093cffab461870b19705161f65b8f27e38041fd0ded773bf0ebc799" Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.344738 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc485fdda093cffab461870b19705161f65b8f27e38041fd0ded773bf0ebc799"} err="failed to get container status \"fc485fdda093cffab461870b19705161f65b8f27e38041fd0ded773bf0ebc799\": rpc error: code = NotFound desc = could not find container \"fc485fdda093cffab461870b19705161f65b8f27e38041fd0ded773bf0ebc799\": container with ID starting with fc485fdda093cffab461870b19705161f65b8f27e38041fd0ded773bf0ebc799 not found: ID does not exist" Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.344760 4559 scope.go:117] "RemoveContainer" containerID="616b63b0514c438639d975b5dac16185db3e95d1a993bb586c1fd317be9f122e" Nov 23 06:58:28 crc kubenswrapper[4559]: E1123 06:58:28.347666 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"616b63b0514c438639d975b5dac16185db3e95d1a993bb586c1fd317be9f122e\": container with ID starting with 616b63b0514c438639d975b5dac16185db3e95d1a993bb586c1fd317be9f122e not found: ID does not exist" containerID="616b63b0514c438639d975b5dac16185db3e95d1a993bb586c1fd317be9f122e" Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.347714 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"616b63b0514c438639d975b5dac16185db3e95d1a993bb586c1fd317be9f122e"} err="failed to get container status \"616b63b0514c438639d975b5dac16185db3e95d1a993bb586c1fd317be9f122e\": rpc error: code = NotFound desc = could not find container \"616b63b0514c438639d975b5dac16185db3e95d1a993bb586c1fd317be9f122e\": container with ID starting with 616b63b0514c438639d975b5dac16185db3e95d1a993bb586c1fd317be9f122e not found: ID does not exist" Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.347741 4559 scope.go:117] "RemoveContainer" containerID="481c1c0cdb7c672d4a4a53468a8b1360b3efaffa13d3dc3b3b8a200072b6173d" Nov 23 06:58:28 crc kubenswrapper[4559]: E1123 06:58:28.349464 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"481c1c0cdb7c672d4a4a53468a8b1360b3efaffa13d3dc3b3b8a200072b6173d\": container with ID starting with 481c1c0cdb7c672d4a4a53468a8b1360b3efaffa13d3dc3b3b8a200072b6173d not found: ID does not exist" containerID="481c1c0cdb7c672d4a4a53468a8b1360b3efaffa13d3dc3b3b8a200072b6173d" Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.349512 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"481c1c0cdb7c672d4a4a53468a8b1360b3efaffa13d3dc3b3b8a200072b6173d"} err="failed to get container status \"481c1c0cdb7c672d4a4a53468a8b1360b3efaffa13d3dc3b3b8a200072b6173d\": rpc error: code = NotFound desc = could not find container \"481c1c0cdb7c672d4a4a53468a8b1360b3efaffa13d3dc3b3b8a200072b6173d\": container with ID starting with 481c1c0cdb7c672d4a4a53468a8b1360b3efaffa13d3dc3b3b8a200072b6173d not found: ID does not exist" Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.423615 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-c1ff-account-create-ngpwh"] Nov 23 06:58:28 crc kubenswrapper[4559]: W1123 06:58:28.427044 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda81225a3_0928_42f9_bdc8_22375a649775.slice/crio-f892ffd53391e5848c45ed42327bfdb7b314ae235ec17a358e903bcd848ffeec WatchSource:0}: Error finding container f892ffd53391e5848c45ed42327bfdb7b314ae235ec17a358e903bcd848ffeec: Status 404 returned error can't find the container with id f892ffd53391e5848c45ed42327bfdb7b314ae235ec17a358e903bcd848ffeec Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.606976 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-556xz"] Nov 23 06:58:28 crc kubenswrapper[4559]: W1123 06:58:28.612428 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf26678a1_0d1d_40c6_9eef_15f8ff08caab.slice/crio-66265c28d00ba52931637cc04e686bda4382ea24929caad197254e430f5a764f WatchSource:0}: Error finding container 66265c28d00ba52931637cc04e686bda4382ea24929caad197254e430f5a764f: Status 404 returned error can't find the container with id 66265c28d00ba52931637cc04e686bda4382ea24929caad197254e430f5a764f Nov 23 06:58:28 crc kubenswrapper[4559]: I1123 06:58:28.749799 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9b47-account-create-br9bd"] Nov 23 06:58:28 crc kubenswrapper[4559]: W1123 06:58:28.753095 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2efe0b7a_32c9_40b6_a933_014c9561b25e.slice/crio-2e7aa84d5c1044e39ce0baef44aaec9ff587f071cbf7830e2b6741a0bd52ec0d WatchSource:0}: Error finding container 2e7aa84d5c1044e39ce0baef44aaec9ff587f071cbf7830e2b6741a0bd52ec0d: Status 404 returned error can't find the container with id 2e7aa84d5c1044e39ce0baef44aaec9ff587f071cbf7830e2b6741a0bd52ec0d Nov 23 06:58:29 crc kubenswrapper[4559]: I1123 06:58:29.260928 4559 generic.go:334] "Generic (PLEG): container finished" podID="2efe0b7a-32c9-40b6-a933-014c9561b25e" containerID="670f508ab528337d97e501177b1ee9f4694a39f89314fc6752425ecda77b578a" exitCode=0 Nov 23 06:58:29 crc kubenswrapper[4559]: I1123 06:58:29.261022 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9b47-account-create-br9bd" event={"ID":"2efe0b7a-32c9-40b6-a933-014c9561b25e","Type":"ContainerDied","Data":"670f508ab528337d97e501177b1ee9f4694a39f89314fc6752425ecda77b578a"} Nov 23 06:58:29 crc kubenswrapper[4559]: I1123 06:58:29.261073 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9b47-account-create-br9bd" event={"ID":"2efe0b7a-32c9-40b6-a933-014c9561b25e","Type":"ContainerStarted","Data":"2e7aa84d5c1044e39ce0baef44aaec9ff587f071cbf7830e2b6741a0bd52ec0d"} Nov 23 06:58:29 crc kubenswrapper[4559]: I1123 06:58:29.263523 4559 generic.go:334] "Generic (PLEG): container finished" podID="f26678a1-0d1d-40c6-9eef-15f8ff08caab" containerID="447e3db48c4b79fa32cad92f618795df76a33cfc391fb7dadf972e3631cb3360" exitCode=0 Nov 23 06:58:29 crc kubenswrapper[4559]: I1123 06:58:29.263570 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-556xz" event={"ID":"f26678a1-0d1d-40c6-9eef-15f8ff08caab","Type":"ContainerDied","Data":"447e3db48c4b79fa32cad92f618795df76a33cfc391fb7dadf972e3631cb3360"} Nov 23 06:58:29 crc kubenswrapper[4559]: I1123 06:58:29.263589 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-556xz" event={"ID":"f26678a1-0d1d-40c6-9eef-15f8ff08caab","Type":"ContainerStarted","Data":"66265c28d00ba52931637cc04e686bda4382ea24929caad197254e430f5a764f"} Nov 23 06:58:29 crc kubenswrapper[4559]: I1123 06:58:29.264962 4559 generic.go:334] "Generic (PLEG): container finished" podID="de1904f5-a9a6-4bc2-b106-71cbc351650b" containerID="50e879df85b3e5d509462b0d8abd7c9217e25644c9e956a57311ed210f16d4c2" exitCode=0 Nov 23 06:58:29 crc kubenswrapper[4559]: I1123 06:58:29.265057 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-48ksj" event={"ID":"de1904f5-a9a6-4bc2-b106-71cbc351650b","Type":"ContainerDied","Data":"50e879df85b3e5d509462b0d8abd7c9217e25644c9e956a57311ed210f16d4c2"} Nov 23 06:58:29 crc kubenswrapper[4559]: I1123 06:58:29.266526 4559 generic.go:334] "Generic (PLEG): container finished" podID="a81225a3-0928-42f9-bdc8-22375a649775" containerID="48f592e4030e6b3434d490cb8ff49b0778303be19fcee13d4246eecff5a75668" exitCode=0 Nov 23 06:58:29 crc kubenswrapper[4559]: I1123 06:58:29.266589 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c1ff-account-create-ngpwh" event={"ID":"a81225a3-0928-42f9-bdc8-22375a649775","Type":"ContainerDied","Data":"48f592e4030e6b3434d490cb8ff49b0778303be19fcee13d4246eecff5a75668"} Nov 23 06:58:29 crc kubenswrapper[4559]: I1123 06:58:29.266622 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c1ff-account-create-ngpwh" event={"ID":"a81225a3-0928-42f9-bdc8-22375a649775","Type":"ContainerStarted","Data":"f892ffd53391e5848c45ed42327bfdb7b314ae235ec17a358e903bcd848ffeec"} Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.281657 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56c603bf-2cb8-4c83-b11f-7794dcc0114a" path="/var/lib/kubelet/pods/56c603bf-2cb8-4c83-b11f-7794dcc0114a/volumes" Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.625751 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-556xz" Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.656154 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q52n2\" (UniqueName: \"kubernetes.io/projected/f26678a1-0d1d-40c6-9eef-15f8ff08caab-kube-api-access-q52n2\") pod \"f26678a1-0d1d-40c6-9eef-15f8ff08caab\" (UID: \"f26678a1-0d1d-40c6-9eef-15f8ff08caab\") " Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.656309 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f26678a1-0d1d-40c6-9eef-15f8ff08caab-operator-scripts\") pod \"f26678a1-0d1d-40c6-9eef-15f8ff08caab\" (UID: \"f26678a1-0d1d-40c6-9eef-15f8ff08caab\") " Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.656792 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f26678a1-0d1d-40c6-9eef-15f8ff08caab-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f26678a1-0d1d-40c6-9eef-15f8ff08caab" (UID: "f26678a1-0d1d-40c6-9eef-15f8ff08caab"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.661043 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f26678a1-0d1d-40c6-9eef-15f8ff08caab-kube-api-access-q52n2" (OuterVolumeSpecName: "kube-api-access-q52n2") pod "f26678a1-0d1d-40c6-9eef-15f8ff08caab" (UID: "f26678a1-0d1d-40c6-9eef-15f8ff08caab"). InnerVolumeSpecName "kube-api-access-q52n2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.703979 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c1ff-account-create-ngpwh" Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.708476 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9b47-account-create-br9bd" Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.712061 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-48ksj" Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.757801 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2efe0b7a-32c9-40b6-a933-014c9561b25e-operator-scripts\") pod \"2efe0b7a-32c9-40b6-a933-014c9561b25e\" (UID: \"2efe0b7a-32c9-40b6-a933-014c9561b25e\") " Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.758140 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a81225a3-0928-42f9-bdc8-22375a649775-operator-scripts\") pod \"a81225a3-0928-42f9-bdc8-22375a649775\" (UID: \"a81225a3-0928-42f9-bdc8-22375a649775\") " Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.758245 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2efe0b7a-32c9-40b6-a933-014c9561b25e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2efe0b7a-32c9-40b6-a933-014c9561b25e" (UID: "2efe0b7a-32c9-40b6-a933-014c9561b25e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.758400 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkcvg\" (UniqueName: \"kubernetes.io/projected/a81225a3-0928-42f9-bdc8-22375a649775-kube-api-access-mkcvg\") pod \"a81225a3-0928-42f9-bdc8-22375a649775\" (UID: \"a81225a3-0928-42f9-bdc8-22375a649775\") " Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.758803 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmcln\" (UniqueName: \"kubernetes.io/projected/2efe0b7a-32c9-40b6-a933-014c9561b25e-kube-api-access-qmcln\") pod \"2efe0b7a-32c9-40b6-a933-014c9561b25e\" (UID: \"2efe0b7a-32c9-40b6-a933-014c9561b25e\") " Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.758982 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdhx9\" (UniqueName: \"kubernetes.io/projected/de1904f5-a9a6-4bc2-b106-71cbc351650b-kube-api-access-wdhx9\") pod \"de1904f5-a9a6-4bc2-b106-71cbc351650b\" (UID: \"de1904f5-a9a6-4bc2-b106-71cbc351650b\") " Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.759093 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de1904f5-a9a6-4bc2-b106-71cbc351650b-operator-scripts\") pod \"de1904f5-a9a6-4bc2-b106-71cbc351650b\" (UID: \"de1904f5-a9a6-4bc2-b106-71cbc351650b\") " Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.758618 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a81225a3-0928-42f9-bdc8-22375a649775-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a81225a3-0928-42f9-bdc8-22375a649775" (UID: "a81225a3-0928-42f9-bdc8-22375a649775"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.759549 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de1904f5-a9a6-4bc2-b106-71cbc351650b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "de1904f5-a9a6-4bc2-b106-71cbc351650b" (UID: "de1904f5-a9a6-4bc2-b106-71cbc351650b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.759988 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q52n2\" (UniqueName: \"kubernetes.io/projected/f26678a1-0d1d-40c6-9eef-15f8ff08caab-kube-api-access-q52n2\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.760067 4559 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de1904f5-a9a6-4bc2-b106-71cbc351650b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.760123 4559 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2efe0b7a-32c9-40b6-a933-014c9561b25e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.760171 4559 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f26678a1-0d1d-40c6-9eef-15f8ff08caab-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.760217 4559 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a81225a3-0928-42f9-bdc8-22375a649775-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.761383 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2efe0b7a-32c9-40b6-a933-014c9561b25e-kube-api-access-qmcln" (OuterVolumeSpecName: "kube-api-access-qmcln") pod "2efe0b7a-32c9-40b6-a933-014c9561b25e" (UID: "2efe0b7a-32c9-40b6-a933-014c9561b25e"). InnerVolumeSpecName "kube-api-access-qmcln". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.763379 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de1904f5-a9a6-4bc2-b106-71cbc351650b-kube-api-access-wdhx9" (OuterVolumeSpecName: "kube-api-access-wdhx9") pod "de1904f5-a9a6-4bc2-b106-71cbc351650b" (UID: "de1904f5-a9a6-4bc2-b106-71cbc351650b"). InnerVolumeSpecName "kube-api-access-wdhx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.763423 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a81225a3-0928-42f9-bdc8-22375a649775-kube-api-access-mkcvg" (OuterVolumeSpecName: "kube-api-access-mkcvg") pod "a81225a3-0928-42f9-bdc8-22375a649775" (UID: "a81225a3-0928-42f9-bdc8-22375a649775"). InnerVolumeSpecName "kube-api-access-mkcvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.861793 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmcln\" (UniqueName: \"kubernetes.io/projected/2efe0b7a-32c9-40b6-a933-014c9561b25e-kube-api-access-qmcln\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.861824 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdhx9\" (UniqueName: \"kubernetes.io/projected/de1904f5-a9a6-4bc2-b106-71cbc351650b-kube-api-access-wdhx9\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:30 crc kubenswrapper[4559]: I1123 06:58:30.861835 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkcvg\" (UniqueName: \"kubernetes.io/projected/a81225a3-0928-42f9-bdc8-22375a649775-kube-api-access-mkcvg\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:31 crc kubenswrapper[4559]: I1123 06:58:31.279497 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-48ksj" Nov 23 06:58:31 crc kubenswrapper[4559]: I1123 06:58:31.279516 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-48ksj" event={"ID":"de1904f5-a9a6-4bc2-b106-71cbc351650b","Type":"ContainerDied","Data":"f9e73b0ab841045907311e0cefaa212dbc9e278213cee9bd31516be4bf282c91"} Nov 23 06:58:31 crc kubenswrapper[4559]: I1123 06:58:31.279559 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9e73b0ab841045907311e0cefaa212dbc9e278213cee9bd31516be4bf282c91" Nov 23 06:58:31 crc kubenswrapper[4559]: I1123 06:58:31.281669 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c1ff-account-create-ngpwh" event={"ID":"a81225a3-0928-42f9-bdc8-22375a649775","Type":"ContainerDied","Data":"f892ffd53391e5848c45ed42327bfdb7b314ae235ec17a358e903bcd848ffeec"} Nov 23 06:58:31 crc kubenswrapper[4559]: I1123 06:58:31.281705 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f892ffd53391e5848c45ed42327bfdb7b314ae235ec17a358e903bcd848ffeec" Nov 23 06:58:31 crc kubenswrapper[4559]: I1123 06:58:31.281692 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c1ff-account-create-ngpwh" Nov 23 06:58:31 crc kubenswrapper[4559]: I1123 06:58:31.282943 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9b47-account-create-br9bd" event={"ID":"2efe0b7a-32c9-40b6-a933-014c9561b25e","Type":"ContainerDied","Data":"2e7aa84d5c1044e39ce0baef44aaec9ff587f071cbf7830e2b6741a0bd52ec0d"} Nov 23 06:58:31 crc kubenswrapper[4559]: I1123 06:58:31.282973 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e7aa84d5c1044e39ce0baef44aaec9ff587f071cbf7830e2b6741a0bd52ec0d" Nov 23 06:58:31 crc kubenswrapper[4559]: I1123 06:58:31.283015 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9b47-account-create-br9bd" Nov 23 06:58:31 crc kubenswrapper[4559]: I1123 06:58:31.285767 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-556xz" event={"ID":"f26678a1-0d1d-40c6-9eef-15f8ff08caab","Type":"ContainerDied","Data":"66265c28d00ba52931637cc04e686bda4382ea24929caad197254e430f5a764f"} Nov 23 06:58:31 crc kubenswrapper[4559]: I1123 06:58:31.285806 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66265c28d00ba52931637cc04e686bda4382ea24929caad197254e430f5a764f" Nov 23 06:58:31 crc kubenswrapper[4559]: I1123 06:58:31.285857 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-556xz" Nov 23 06:58:32 crc kubenswrapper[4559]: I1123 06:58:32.989852 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b8455895f-w72ch" Nov 23 06:58:32 crc kubenswrapper[4559]: I1123 06:58:32.995616 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.030393 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-759c6cc4df-s47tx"] Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.030606 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-759c6cc4df-s47tx" podUID="0e6213f8-9948-47e9-88ac-7d7bdaddac26" containerName="dnsmasq-dns" containerID="cri-o://49c281f5966b944fb67c9b11ed9bf54b265407038016add98feccbd852cc8661" gracePeriod=10 Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.144227 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-njl4x"] Nov 23 06:58:33 crc kubenswrapper[4559]: E1123 06:58:33.144565 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de1904f5-a9a6-4bc2-b106-71cbc351650b" containerName="mariadb-database-create" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.144586 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="de1904f5-a9a6-4bc2-b106-71cbc351650b" containerName="mariadb-database-create" Nov 23 06:58:33 crc kubenswrapper[4559]: E1123 06:58:33.144597 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f26678a1-0d1d-40c6-9eef-15f8ff08caab" containerName="mariadb-database-create" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.144603 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="f26678a1-0d1d-40c6-9eef-15f8ff08caab" containerName="mariadb-database-create" Nov 23 06:58:33 crc kubenswrapper[4559]: E1123 06:58:33.144620 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a81225a3-0928-42f9-bdc8-22375a649775" containerName="mariadb-account-create" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.144626 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="a81225a3-0928-42f9-bdc8-22375a649775" containerName="mariadb-account-create" Nov 23 06:58:33 crc kubenswrapper[4559]: E1123 06:58:33.144679 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2efe0b7a-32c9-40b6-a933-014c9561b25e" containerName="mariadb-account-create" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.144685 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="2efe0b7a-32c9-40b6-a933-014c9561b25e" containerName="mariadb-account-create" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.147762 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="a81225a3-0928-42f9-bdc8-22375a649775" containerName="mariadb-account-create" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.147791 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="f26678a1-0d1d-40c6-9eef-15f8ff08caab" containerName="mariadb-database-create" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.147817 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="de1904f5-a9a6-4bc2-b106-71cbc351650b" containerName="mariadb-database-create" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.147843 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="2efe0b7a-32c9-40b6-a933-014c9561b25e" containerName="mariadb-account-create" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.148529 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-njl4x" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.174248 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-njl4x"] Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.198050 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhpwb\" (UniqueName: \"kubernetes.io/projected/f5018f2c-cf19-4991-a3ad-5909128d9de9-kube-api-access-rhpwb\") pod \"glance-db-create-njl4x\" (UID: \"f5018f2c-cf19-4991-a3ad-5909128d9de9\") " pod="openstack/glance-db-create-njl4x" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.198171 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5018f2c-cf19-4991-a3ad-5909128d9de9-operator-scripts\") pod \"glance-db-create-njl4x\" (UID: \"f5018f2c-cf19-4991-a3ad-5909128d9de9\") " pod="openstack/glance-db-create-njl4x" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.243360 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-7b13-account-create-mzr4r"] Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.244602 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-7b13-account-create-mzr4r" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.246674 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.257809 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-7b13-account-create-mzr4r"] Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.299590 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37737b2f-1fb8-4028-a7a2-134fb6e0b991-operator-scripts\") pod \"glance-7b13-account-create-mzr4r\" (UID: \"37737b2f-1fb8-4028-a7a2-134fb6e0b991\") " pod="openstack/glance-7b13-account-create-mzr4r" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.299801 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5018f2c-cf19-4991-a3ad-5909128d9de9-operator-scripts\") pod \"glance-db-create-njl4x\" (UID: \"f5018f2c-cf19-4991-a3ad-5909128d9de9\") " pod="openstack/glance-db-create-njl4x" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.299848 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqbtw\" (UniqueName: \"kubernetes.io/projected/37737b2f-1fb8-4028-a7a2-134fb6e0b991-kube-api-access-mqbtw\") pod \"glance-7b13-account-create-mzr4r\" (UID: \"37737b2f-1fb8-4028-a7a2-134fb6e0b991\") " pod="openstack/glance-7b13-account-create-mzr4r" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.300105 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhpwb\" (UniqueName: \"kubernetes.io/projected/f5018f2c-cf19-4991-a3ad-5909128d9de9-kube-api-access-rhpwb\") pod \"glance-db-create-njl4x\" (UID: \"f5018f2c-cf19-4991-a3ad-5909128d9de9\") " pod="openstack/glance-db-create-njl4x" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.300712 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5018f2c-cf19-4991-a3ad-5909128d9de9-operator-scripts\") pod \"glance-db-create-njl4x\" (UID: \"f5018f2c-cf19-4991-a3ad-5909128d9de9\") " pod="openstack/glance-db-create-njl4x" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.301897 4559 generic.go:334] "Generic (PLEG): container finished" podID="0e6213f8-9948-47e9-88ac-7d7bdaddac26" containerID="49c281f5966b944fb67c9b11ed9bf54b265407038016add98feccbd852cc8661" exitCode=0 Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.301929 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-759c6cc4df-s47tx" event={"ID":"0e6213f8-9948-47e9-88ac-7d7bdaddac26","Type":"ContainerDied","Data":"49c281f5966b944fb67c9b11ed9bf54b265407038016add98feccbd852cc8661"} Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.317852 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhpwb\" (UniqueName: \"kubernetes.io/projected/f5018f2c-cf19-4991-a3ad-5909128d9de9-kube-api-access-rhpwb\") pod \"glance-db-create-njl4x\" (UID: \"f5018f2c-cf19-4991-a3ad-5909128d9de9\") " pod="openstack/glance-db-create-njl4x" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.401795 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37737b2f-1fb8-4028-a7a2-134fb6e0b991-operator-scripts\") pod \"glance-7b13-account-create-mzr4r\" (UID: \"37737b2f-1fb8-4028-a7a2-134fb6e0b991\") " pod="openstack/glance-7b13-account-create-mzr4r" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.402020 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqbtw\" (UniqueName: \"kubernetes.io/projected/37737b2f-1fb8-4028-a7a2-134fb6e0b991-kube-api-access-mqbtw\") pod \"glance-7b13-account-create-mzr4r\" (UID: \"37737b2f-1fb8-4028-a7a2-134fb6e0b991\") " pod="openstack/glance-7b13-account-create-mzr4r" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.402597 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37737b2f-1fb8-4028-a7a2-134fb6e0b991-operator-scripts\") pod \"glance-7b13-account-create-mzr4r\" (UID: \"37737b2f-1fb8-4028-a7a2-134fb6e0b991\") " pod="openstack/glance-7b13-account-create-mzr4r" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.417453 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqbtw\" (UniqueName: \"kubernetes.io/projected/37737b2f-1fb8-4028-a7a2-134fb6e0b991-kube-api-access-mqbtw\") pod \"glance-7b13-account-create-mzr4r\" (UID: \"37737b2f-1fb8-4028-a7a2-134fb6e0b991\") " pod="openstack/glance-7b13-account-create-mzr4r" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.460769 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-759c6cc4df-s47tx" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.486683 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-njl4x" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.503490 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e6213f8-9948-47e9-88ac-7d7bdaddac26-dns-svc\") pod \"0e6213f8-9948-47e9-88ac-7d7bdaddac26\" (UID: \"0e6213f8-9948-47e9-88ac-7d7bdaddac26\") " Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.503555 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e6213f8-9948-47e9-88ac-7d7bdaddac26-config\") pod \"0e6213f8-9948-47e9-88ac-7d7bdaddac26\" (UID: \"0e6213f8-9948-47e9-88ac-7d7bdaddac26\") " Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.503726 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6jq2\" (UniqueName: \"kubernetes.io/projected/0e6213f8-9948-47e9-88ac-7d7bdaddac26-kube-api-access-q6jq2\") pod \"0e6213f8-9948-47e9-88ac-7d7bdaddac26\" (UID: \"0e6213f8-9948-47e9-88ac-7d7bdaddac26\") " Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.507665 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e6213f8-9948-47e9-88ac-7d7bdaddac26-kube-api-access-q6jq2" (OuterVolumeSpecName: "kube-api-access-q6jq2") pod "0e6213f8-9948-47e9-88ac-7d7bdaddac26" (UID: "0e6213f8-9948-47e9-88ac-7d7bdaddac26"). InnerVolumeSpecName "kube-api-access-q6jq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.534256 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e6213f8-9948-47e9-88ac-7d7bdaddac26-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0e6213f8-9948-47e9-88ac-7d7bdaddac26" (UID: "0e6213f8-9948-47e9-88ac-7d7bdaddac26"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.539997 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e6213f8-9948-47e9-88ac-7d7bdaddac26-config" (OuterVolumeSpecName: "config") pod "0e6213f8-9948-47e9-88ac-7d7bdaddac26" (UID: "0e6213f8-9948-47e9-88ac-7d7bdaddac26"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.559103 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-7b13-account-create-mzr4r" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.607977 4559 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e6213f8-9948-47e9-88ac-7d7bdaddac26-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.608005 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e6213f8-9948-47e9-88ac-7d7bdaddac26-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.608015 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6jq2\" (UniqueName: \"kubernetes.io/projected/0e6213f8-9948-47e9-88ac-7d7bdaddac26-kube-api-access-q6jq2\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.873461 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-njl4x"] Nov 23 06:58:33 crc kubenswrapper[4559]: W1123 06:58:33.875094 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5018f2c_cf19_4991_a3ad_5909128d9de9.slice/crio-8523980821a93db9f3923035c931451b7258c2b91fac39856b14e2eb86f3090b WatchSource:0}: Error finding container 8523980821a93db9f3923035c931451b7258c2b91fac39856b14e2eb86f3090b: Status 404 returned error can't find the container with id 8523980821a93db9f3923035c931451b7258c2b91fac39856b14e2eb86f3090b Nov 23 06:58:33 crc kubenswrapper[4559]: I1123 06:58:33.949160 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-7b13-account-create-mzr4r"] Nov 23 06:58:33 crc kubenswrapper[4559]: W1123 06:58:33.957626 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37737b2f_1fb8_4028_a7a2_134fb6e0b991.slice/crio-302ef8126e5bb5dff12cc2266858b0634b10102c4f1debf79cd2ad0d9da553d0 WatchSource:0}: Error finding container 302ef8126e5bb5dff12cc2266858b0634b10102c4f1debf79cd2ad0d9da553d0: Status 404 returned error can't find the container with id 302ef8126e5bb5dff12cc2266858b0634b10102c4f1debf79cd2ad0d9da553d0 Nov 23 06:58:34 crc kubenswrapper[4559]: I1123 06:58:34.311815 4559 generic.go:334] "Generic (PLEG): container finished" podID="f5018f2c-cf19-4991-a3ad-5909128d9de9" containerID="f3cbd5748e3000217e906faa8173d52b20792663728a89a32049aba6b07e3e6b" exitCode=0 Nov 23 06:58:34 crc kubenswrapper[4559]: I1123 06:58:34.311860 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-njl4x" event={"ID":"f5018f2c-cf19-4991-a3ad-5909128d9de9","Type":"ContainerDied","Data":"f3cbd5748e3000217e906faa8173d52b20792663728a89a32049aba6b07e3e6b"} Nov 23 06:58:34 crc kubenswrapper[4559]: I1123 06:58:34.311930 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-njl4x" event={"ID":"f5018f2c-cf19-4991-a3ad-5909128d9de9","Type":"ContainerStarted","Data":"8523980821a93db9f3923035c931451b7258c2b91fac39856b14e2eb86f3090b"} Nov 23 06:58:34 crc kubenswrapper[4559]: I1123 06:58:34.313778 4559 generic.go:334] "Generic (PLEG): container finished" podID="37737b2f-1fb8-4028-a7a2-134fb6e0b991" containerID="dc01e112a1fab31fb727f9b68a2057b14624fbf2f00723615b543be9cc67de1e" exitCode=0 Nov 23 06:58:34 crc kubenswrapper[4559]: I1123 06:58:34.313828 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-7b13-account-create-mzr4r" event={"ID":"37737b2f-1fb8-4028-a7a2-134fb6e0b991","Type":"ContainerDied","Data":"dc01e112a1fab31fb727f9b68a2057b14624fbf2f00723615b543be9cc67de1e"} Nov 23 06:58:34 crc kubenswrapper[4559]: I1123 06:58:34.313861 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-7b13-account-create-mzr4r" event={"ID":"37737b2f-1fb8-4028-a7a2-134fb6e0b991","Type":"ContainerStarted","Data":"302ef8126e5bb5dff12cc2266858b0634b10102c4f1debf79cd2ad0d9da553d0"} Nov 23 06:58:34 crc kubenswrapper[4559]: I1123 06:58:34.316071 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-759c6cc4df-s47tx" event={"ID":"0e6213f8-9948-47e9-88ac-7d7bdaddac26","Type":"ContainerDied","Data":"117c6950fa5e853b90db417edea073284ebf5b1fa2b4d2bd374fb78495526e8f"} Nov 23 06:58:34 crc kubenswrapper[4559]: I1123 06:58:34.316116 4559 scope.go:117] "RemoveContainer" containerID="49c281f5966b944fb67c9b11ed9bf54b265407038016add98feccbd852cc8661" Nov 23 06:58:34 crc kubenswrapper[4559]: I1123 06:58:34.316211 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-759c6cc4df-s47tx" Nov 23 06:58:34 crc kubenswrapper[4559]: I1123 06:58:34.355341 4559 scope.go:117] "RemoveContainer" containerID="ff867a700a523808a42c0dc55d10a2a7c544c16dc2e540c82cc4bdd7eeaae817" Nov 23 06:58:34 crc kubenswrapper[4559]: I1123 06:58:34.359475 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-759c6cc4df-s47tx"] Nov 23 06:58:34 crc kubenswrapper[4559]: I1123 06:58:34.365396 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-759c6cc4df-s47tx"] Nov 23 06:58:35 crc kubenswrapper[4559]: I1123 06:58:35.648670 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-njl4x" Nov 23 06:58:35 crc kubenswrapper[4559]: I1123 06:58:35.653384 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-7b13-account-create-mzr4r" Nov 23 06:58:35 crc kubenswrapper[4559]: I1123 06:58:35.738968 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37737b2f-1fb8-4028-a7a2-134fb6e0b991-operator-scripts\") pod \"37737b2f-1fb8-4028-a7a2-134fb6e0b991\" (UID: \"37737b2f-1fb8-4028-a7a2-134fb6e0b991\") " Nov 23 06:58:35 crc kubenswrapper[4559]: I1123 06:58:35.739016 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhpwb\" (UniqueName: \"kubernetes.io/projected/f5018f2c-cf19-4991-a3ad-5909128d9de9-kube-api-access-rhpwb\") pod \"f5018f2c-cf19-4991-a3ad-5909128d9de9\" (UID: \"f5018f2c-cf19-4991-a3ad-5909128d9de9\") " Nov 23 06:58:35 crc kubenswrapper[4559]: I1123 06:58:35.739057 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5018f2c-cf19-4991-a3ad-5909128d9de9-operator-scripts\") pod \"f5018f2c-cf19-4991-a3ad-5909128d9de9\" (UID: \"f5018f2c-cf19-4991-a3ad-5909128d9de9\") " Nov 23 06:58:35 crc kubenswrapper[4559]: I1123 06:58:35.739142 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqbtw\" (UniqueName: \"kubernetes.io/projected/37737b2f-1fb8-4028-a7a2-134fb6e0b991-kube-api-access-mqbtw\") pod \"37737b2f-1fb8-4028-a7a2-134fb6e0b991\" (UID: \"37737b2f-1fb8-4028-a7a2-134fb6e0b991\") " Nov 23 06:58:35 crc kubenswrapper[4559]: I1123 06:58:35.739752 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37737b2f-1fb8-4028-a7a2-134fb6e0b991-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "37737b2f-1fb8-4028-a7a2-134fb6e0b991" (UID: "37737b2f-1fb8-4028-a7a2-134fb6e0b991"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:58:35 crc kubenswrapper[4559]: I1123 06:58:35.740055 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5018f2c-cf19-4991-a3ad-5909128d9de9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f5018f2c-cf19-4991-a3ad-5909128d9de9" (UID: "f5018f2c-cf19-4991-a3ad-5909128d9de9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:58:35 crc kubenswrapper[4559]: I1123 06:58:35.743860 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37737b2f-1fb8-4028-a7a2-134fb6e0b991-kube-api-access-mqbtw" (OuterVolumeSpecName: "kube-api-access-mqbtw") pod "37737b2f-1fb8-4028-a7a2-134fb6e0b991" (UID: "37737b2f-1fb8-4028-a7a2-134fb6e0b991"). InnerVolumeSpecName "kube-api-access-mqbtw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:58:35 crc kubenswrapper[4559]: I1123 06:58:35.744213 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5018f2c-cf19-4991-a3ad-5909128d9de9-kube-api-access-rhpwb" (OuterVolumeSpecName: "kube-api-access-rhpwb") pod "f5018f2c-cf19-4991-a3ad-5909128d9de9" (UID: "f5018f2c-cf19-4991-a3ad-5909128d9de9"). InnerVolumeSpecName "kube-api-access-rhpwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:58:35 crc kubenswrapper[4559]: I1123 06:58:35.840573 4559 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37737b2f-1fb8-4028-a7a2-134fb6e0b991-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:35 crc kubenswrapper[4559]: I1123 06:58:35.840596 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhpwb\" (UniqueName: \"kubernetes.io/projected/f5018f2c-cf19-4991-a3ad-5909128d9de9-kube-api-access-rhpwb\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:35 crc kubenswrapper[4559]: I1123 06:58:35.840607 4559 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5018f2c-cf19-4991-a3ad-5909128d9de9-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:35 crc kubenswrapper[4559]: I1123 06:58:35.840616 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqbtw\" (UniqueName: \"kubernetes.io/projected/37737b2f-1fb8-4028-a7a2-134fb6e0b991-kube-api-access-mqbtw\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:36 crc kubenswrapper[4559]: I1123 06:58:36.284439 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e6213f8-9948-47e9-88ac-7d7bdaddac26" path="/var/lib/kubelet/pods/0e6213f8-9948-47e9-88ac-7d7bdaddac26/volumes" Nov 23 06:58:36 crc kubenswrapper[4559]: I1123 06:58:36.334467 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-7b13-account-create-mzr4r" event={"ID":"37737b2f-1fb8-4028-a7a2-134fb6e0b991","Type":"ContainerDied","Data":"302ef8126e5bb5dff12cc2266858b0634b10102c4f1debf79cd2ad0d9da553d0"} Nov 23 06:58:36 crc kubenswrapper[4559]: I1123 06:58:36.334725 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="302ef8126e5bb5dff12cc2266858b0634b10102c4f1debf79cd2ad0d9da553d0" Nov 23 06:58:36 crc kubenswrapper[4559]: I1123 06:58:36.334477 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-7b13-account-create-mzr4r" Nov 23 06:58:36 crc kubenswrapper[4559]: I1123 06:58:36.336209 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-njl4x" event={"ID":"f5018f2c-cf19-4991-a3ad-5909128d9de9","Type":"ContainerDied","Data":"8523980821a93db9f3923035c931451b7258c2b91fac39856b14e2eb86f3090b"} Nov 23 06:58:36 crc kubenswrapper[4559]: I1123 06:58:36.336250 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8523980821a93db9f3923035c931451b7258c2b91fac39856b14e2eb86f3090b" Nov 23 06:58:36 crc kubenswrapper[4559]: I1123 06:58:36.336231 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-njl4x" Nov 23 06:58:36 crc kubenswrapper[4559]: E1123 06:58:36.357973 4559 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37737b2f_1fb8_4028_a7a2_134fb6e0b991.slice\": RecentStats: unable to find data in memory cache]" Nov 23 06:58:38 crc kubenswrapper[4559]: I1123 06:58:38.026964 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 23 06:58:38 crc kubenswrapper[4559]: I1123 06:58:38.467077 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-dwbhb"] Nov 23 06:58:38 crc kubenswrapper[4559]: E1123 06:58:38.467462 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37737b2f-1fb8-4028-a7a2-134fb6e0b991" containerName="mariadb-account-create" Nov 23 06:58:38 crc kubenswrapper[4559]: I1123 06:58:38.467483 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="37737b2f-1fb8-4028-a7a2-134fb6e0b991" containerName="mariadb-account-create" Nov 23 06:58:38 crc kubenswrapper[4559]: E1123 06:58:38.467499 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e6213f8-9948-47e9-88ac-7d7bdaddac26" containerName="dnsmasq-dns" Nov 23 06:58:38 crc kubenswrapper[4559]: I1123 06:58:38.467507 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e6213f8-9948-47e9-88ac-7d7bdaddac26" containerName="dnsmasq-dns" Nov 23 06:58:38 crc kubenswrapper[4559]: E1123 06:58:38.467523 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5018f2c-cf19-4991-a3ad-5909128d9de9" containerName="mariadb-database-create" Nov 23 06:58:38 crc kubenswrapper[4559]: I1123 06:58:38.467529 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5018f2c-cf19-4991-a3ad-5909128d9de9" containerName="mariadb-database-create" Nov 23 06:58:38 crc kubenswrapper[4559]: E1123 06:58:38.467546 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e6213f8-9948-47e9-88ac-7d7bdaddac26" containerName="init" Nov 23 06:58:38 crc kubenswrapper[4559]: I1123 06:58:38.467552 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e6213f8-9948-47e9-88ac-7d7bdaddac26" containerName="init" Nov 23 06:58:38 crc kubenswrapper[4559]: I1123 06:58:38.467741 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5018f2c-cf19-4991-a3ad-5909128d9de9" containerName="mariadb-database-create" Nov 23 06:58:38 crc kubenswrapper[4559]: I1123 06:58:38.467766 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="37737b2f-1fb8-4028-a7a2-134fb6e0b991" containerName="mariadb-account-create" Nov 23 06:58:38 crc kubenswrapper[4559]: I1123 06:58:38.467781 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e6213f8-9948-47e9-88ac-7d7bdaddac26" containerName="dnsmasq-dns" Nov 23 06:58:38 crc kubenswrapper[4559]: I1123 06:58:38.468379 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-dwbhb" Nov 23 06:58:38 crc kubenswrapper[4559]: I1123 06:58:38.470486 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 23 06:58:38 crc kubenswrapper[4559]: I1123 06:58:38.472958 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-dwbhb"] Nov 23 06:58:38 crc kubenswrapper[4559]: I1123 06:58:38.475484 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-tj89j" Nov 23 06:58:38 crc kubenswrapper[4559]: I1123 06:58:38.575267 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58t7r\" (UniqueName: \"kubernetes.io/projected/a5cc366d-e466-476d-a318-1582a4de4a93-kube-api-access-58t7r\") pod \"glance-db-sync-dwbhb\" (UID: \"a5cc366d-e466-476d-a318-1582a4de4a93\") " pod="openstack/glance-db-sync-dwbhb" Nov 23 06:58:38 crc kubenswrapper[4559]: I1123 06:58:38.575373 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5cc366d-e466-476d-a318-1582a4de4a93-config-data\") pod \"glance-db-sync-dwbhb\" (UID: \"a5cc366d-e466-476d-a318-1582a4de4a93\") " pod="openstack/glance-db-sync-dwbhb" Nov 23 06:58:38 crc kubenswrapper[4559]: I1123 06:58:38.575431 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cc366d-e466-476d-a318-1582a4de4a93-combined-ca-bundle\") pod \"glance-db-sync-dwbhb\" (UID: \"a5cc366d-e466-476d-a318-1582a4de4a93\") " pod="openstack/glance-db-sync-dwbhb" Nov 23 06:58:38 crc kubenswrapper[4559]: I1123 06:58:38.575489 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a5cc366d-e466-476d-a318-1582a4de4a93-db-sync-config-data\") pod \"glance-db-sync-dwbhb\" (UID: \"a5cc366d-e466-476d-a318-1582a4de4a93\") " pod="openstack/glance-db-sync-dwbhb" Nov 23 06:58:38 crc kubenswrapper[4559]: I1123 06:58:38.676756 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5cc366d-e466-476d-a318-1582a4de4a93-config-data\") pod \"glance-db-sync-dwbhb\" (UID: \"a5cc366d-e466-476d-a318-1582a4de4a93\") " pod="openstack/glance-db-sync-dwbhb" Nov 23 06:58:38 crc kubenswrapper[4559]: I1123 06:58:38.676861 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cc366d-e466-476d-a318-1582a4de4a93-combined-ca-bundle\") pod \"glance-db-sync-dwbhb\" (UID: \"a5cc366d-e466-476d-a318-1582a4de4a93\") " pod="openstack/glance-db-sync-dwbhb" Nov 23 06:58:38 crc kubenswrapper[4559]: I1123 06:58:38.676927 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a5cc366d-e466-476d-a318-1582a4de4a93-db-sync-config-data\") pod \"glance-db-sync-dwbhb\" (UID: \"a5cc366d-e466-476d-a318-1582a4de4a93\") " pod="openstack/glance-db-sync-dwbhb" Nov 23 06:58:38 crc kubenswrapper[4559]: I1123 06:58:38.677096 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58t7r\" (UniqueName: \"kubernetes.io/projected/a5cc366d-e466-476d-a318-1582a4de4a93-kube-api-access-58t7r\") pod \"glance-db-sync-dwbhb\" (UID: \"a5cc366d-e466-476d-a318-1582a4de4a93\") " pod="openstack/glance-db-sync-dwbhb" Nov 23 06:58:38 crc kubenswrapper[4559]: I1123 06:58:38.685138 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a5cc366d-e466-476d-a318-1582a4de4a93-db-sync-config-data\") pod \"glance-db-sync-dwbhb\" (UID: \"a5cc366d-e466-476d-a318-1582a4de4a93\") " pod="openstack/glance-db-sync-dwbhb" Nov 23 06:58:38 crc kubenswrapper[4559]: I1123 06:58:38.686185 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cc366d-e466-476d-a318-1582a4de4a93-combined-ca-bundle\") pod \"glance-db-sync-dwbhb\" (UID: \"a5cc366d-e466-476d-a318-1582a4de4a93\") " pod="openstack/glance-db-sync-dwbhb" Nov 23 06:58:38 crc kubenswrapper[4559]: I1123 06:58:38.687062 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5cc366d-e466-476d-a318-1582a4de4a93-config-data\") pod \"glance-db-sync-dwbhb\" (UID: \"a5cc366d-e466-476d-a318-1582a4de4a93\") " pod="openstack/glance-db-sync-dwbhb" Nov 23 06:58:38 crc kubenswrapper[4559]: I1123 06:58:38.707739 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58t7r\" (UniqueName: \"kubernetes.io/projected/a5cc366d-e466-476d-a318-1582a4de4a93-kube-api-access-58t7r\") pod \"glance-db-sync-dwbhb\" (UID: \"a5cc366d-e466-476d-a318-1582a4de4a93\") " pod="openstack/glance-db-sync-dwbhb" Nov 23 06:58:38 crc kubenswrapper[4559]: I1123 06:58:38.782169 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-dwbhb" Nov 23 06:58:39 crc kubenswrapper[4559]: I1123 06:58:39.252949 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-dwbhb"] Nov 23 06:58:39 crc kubenswrapper[4559]: W1123 06:58:39.256328 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5cc366d_e466_476d_a318_1582a4de4a93.slice/crio-79c52b10f05d82764a475ebd15cc13f73d77ad2e6226ec8c24c50873d1710b4f WatchSource:0}: Error finding container 79c52b10f05d82764a475ebd15cc13f73d77ad2e6226ec8c24c50873d1710b4f: Status 404 returned error can't find the container with id 79c52b10f05d82764a475ebd15cc13f73d77ad2e6226ec8c24c50873d1710b4f Nov 23 06:58:39 crc kubenswrapper[4559]: I1123 06:58:39.258680 4559 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 06:58:39 crc kubenswrapper[4559]: I1123 06:58:39.359517 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-dwbhb" event={"ID":"a5cc366d-e466-476d-a318-1582a4de4a93","Type":"ContainerStarted","Data":"79c52b10f05d82764a475ebd15cc13f73d77ad2e6226ec8c24c50873d1710b4f"} Nov 23 06:58:39 crc kubenswrapper[4559]: I1123 06:58:39.428014 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7dfd8c6765-hrs74"] Nov 23 06:58:39 crc kubenswrapper[4559]: I1123 06:58:39.429418 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7dfd8c6765-hrs74" Nov 23 06:58:39 crc kubenswrapper[4559]: I1123 06:58:39.449381 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7dfd8c6765-hrs74"] Nov 23 06:58:39 crc kubenswrapper[4559]: I1123 06:58:39.490465 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35d05808-5864-4a8f-ab67-e4627d8c8770-ovsdbserver-nb\") pod \"dnsmasq-dns-7dfd8c6765-hrs74\" (UID: \"35d05808-5864-4a8f-ab67-e4627d8c8770\") " pod="openstack/dnsmasq-dns-7dfd8c6765-hrs74" Nov 23 06:58:39 crc kubenswrapper[4559]: I1123 06:58:39.490698 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ql2d\" (UniqueName: \"kubernetes.io/projected/35d05808-5864-4a8f-ab67-e4627d8c8770-kube-api-access-5ql2d\") pod \"dnsmasq-dns-7dfd8c6765-hrs74\" (UID: \"35d05808-5864-4a8f-ab67-e4627d8c8770\") " pod="openstack/dnsmasq-dns-7dfd8c6765-hrs74" Nov 23 06:58:39 crc kubenswrapper[4559]: I1123 06:58:39.490741 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35d05808-5864-4a8f-ab67-e4627d8c8770-dns-svc\") pod \"dnsmasq-dns-7dfd8c6765-hrs74\" (UID: \"35d05808-5864-4a8f-ab67-e4627d8c8770\") " pod="openstack/dnsmasq-dns-7dfd8c6765-hrs74" Nov 23 06:58:39 crc kubenswrapper[4559]: I1123 06:58:39.490783 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35d05808-5864-4a8f-ab67-e4627d8c8770-config\") pod \"dnsmasq-dns-7dfd8c6765-hrs74\" (UID: \"35d05808-5864-4a8f-ab67-e4627d8c8770\") " pod="openstack/dnsmasq-dns-7dfd8c6765-hrs74" Nov 23 06:58:39 crc kubenswrapper[4559]: I1123 06:58:39.490820 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35d05808-5864-4a8f-ab67-e4627d8c8770-ovsdbserver-sb\") pod \"dnsmasq-dns-7dfd8c6765-hrs74\" (UID: \"35d05808-5864-4a8f-ab67-e4627d8c8770\") " pod="openstack/dnsmasq-dns-7dfd8c6765-hrs74" Nov 23 06:58:39 crc kubenswrapper[4559]: I1123 06:58:39.592634 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35d05808-5864-4a8f-ab67-e4627d8c8770-ovsdbserver-sb\") pod \"dnsmasq-dns-7dfd8c6765-hrs74\" (UID: \"35d05808-5864-4a8f-ab67-e4627d8c8770\") " pod="openstack/dnsmasq-dns-7dfd8c6765-hrs74" Nov 23 06:58:39 crc kubenswrapper[4559]: I1123 06:58:39.592737 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35d05808-5864-4a8f-ab67-e4627d8c8770-ovsdbserver-nb\") pod \"dnsmasq-dns-7dfd8c6765-hrs74\" (UID: \"35d05808-5864-4a8f-ab67-e4627d8c8770\") " pod="openstack/dnsmasq-dns-7dfd8c6765-hrs74" Nov 23 06:58:39 crc kubenswrapper[4559]: I1123 06:58:39.592969 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ql2d\" (UniqueName: \"kubernetes.io/projected/35d05808-5864-4a8f-ab67-e4627d8c8770-kube-api-access-5ql2d\") pod \"dnsmasq-dns-7dfd8c6765-hrs74\" (UID: \"35d05808-5864-4a8f-ab67-e4627d8c8770\") " pod="openstack/dnsmasq-dns-7dfd8c6765-hrs74" Nov 23 06:58:39 crc kubenswrapper[4559]: I1123 06:58:39.593027 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35d05808-5864-4a8f-ab67-e4627d8c8770-dns-svc\") pod \"dnsmasq-dns-7dfd8c6765-hrs74\" (UID: \"35d05808-5864-4a8f-ab67-e4627d8c8770\") " pod="openstack/dnsmasq-dns-7dfd8c6765-hrs74" Nov 23 06:58:39 crc kubenswrapper[4559]: I1123 06:58:39.593092 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35d05808-5864-4a8f-ab67-e4627d8c8770-config\") pod \"dnsmasq-dns-7dfd8c6765-hrs74\" (UID: \"35d05808-5864-4a8f-ab67-e4627d8c8770\") " pod="openstack/dnsmasq-dns-7dfd8c6765-hrs74" Nov 23 06:58:39 crc kubenswrapper[4559]: I1123 06:58:39.593681 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35d05808-5864-4a8f-ab67-e4627d8c8770-ovsdbserver-sb\") pod \"dnsmasq-dns-7dfd8c6765-hrs74\" (UID: \"35d05808-5864-4a8f-ab67-e4627d8c8770\") " pod="openstack/dnsmasq-dns-7dfd8c6765-hrs74" Nov 23 06:58:39 crc kubenswrapper[4559]: I1123 06:58:39.593830 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35d05808-5864-4a8f-ab67-e4627d8c8770-dns-svc\") pod \"dnsmasq-dns-7dfd8c6765-hrs74\" (UID: \"35d05808-5864-4a8f-ab67-e4627d8c8770\") " pod="openstack/dnsmasq-dns-7dfd8c6765-hrs74" Nov 23 06:58:39 crc kubenswrapper[4559]: I1123 06:58:39.594010 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35d05808-5864-4a8f-ab67-e4627d8c8770-config\") pod \"dnsmasq-dns-7dfd8c6765-hrs74\" (UID: \"35d05808-5864-4a8f-ab67-e4627d8c8770\") " pod="openstack/dnsmasq-dns-7dfd8c6765-hrs74" Nov 23 06:58:39 crc kubenswrapper[4559]: I1123 06:58:39.594356 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35d05808-5864-4a8f-ab67-e4627d8c8770-ovsdbserver-nb\") pod \"dnsmasq-dns-7dfd8c6765-hrs74\" (UID: \"35d05808-5864-4a8f-ab67-e4627d8c8770\") " pod="openstack/dnsmasq-dns-7dfd8c6765-hrs74" Nov 23 06:58:39 crc kubenswrapper[4559]: I1123 06:58:39.609706 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ql2d\" (UniqueName: \"kubernetes.io/projected/35d05808-5864-4a8f-ab67-e4627d8c8770-kube-api-access-5ql2d\") pod \"dnsmasq-dns-7dfd8c6765-hrs74\" (UID: \"35d05808-5864-4a8f-ab67-e4627d8c8770\") " pod="openstack/dnsmasq-dns-7dfd8c6765-hrs74" Nov 23 06:58:39 crc kubenswrapper[4559]: I1123 06:58:39.747624 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7dfd8c6765-hrs74" Nov 23 06:58:40 crc kubenswrapper[4559]: I1123 06:58:40.139542 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7dfd8c6765-hrs74"] Nov 23 06:58:40 crc kubenswrapper[4559]: I1123 06:58:40.369364 4559 generic.go:334] "Generic (PLEG): container finished" podID="35d05808-5864-4a8f-ab67-e4627d8c8770" containerID="2bf22d0a461b496602b5c1dd596999ec1d4b2319de0254ce2fdc3ace79b54bb8" exitCode=0 Nov 23 06:58:40 crc kubenswrapper[4559]: I1123 06:58:40.369796 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7dfd8c6765-hrs74" event={"ID":"35d05808-5864-4a8f-ab67-e4627d8c8770","Type":"ContainerDied","Data":"2bf22d0a461b496602b5c1dd596999ec1d4b2319de0254ce2fdc3ace79b54bb8"} Nov 23 06:58:40 crc kubenswrapper[4559]: I1123 06:58:40.369839 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7dfd8c6765-hrs74" event={"ID":"35d05808-5864-4a8f-ab67-e4627d8c8770","Type":"ContainerStarted","Data":"8a9f56998f257076a74d5200ed3839b2095925a3322ad73d7a6b8996b321f900"} Nov 23 06:58:40 crc kubenswrapper[4559]: I1123 06:58:40.531392 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Nov 23 06:58:40 crc kubenswrapper[4559]: I1123 06:58:40.537496 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 23 06:58:40 crc kubenswrapper[4559]: I1123 06:58:40.539402 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 23 06:58:40 crc kubenswrapper[4559]: I1123 06:58:40.539631 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 23 06:58:40 crc kubenswrapper[4559]: I1123 06:58:40.542803 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-5vsx9" Nov 23 06:58:40 crc kubenswrapper[4559]: I1123 06:58:40.542813 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 23 06:58:40 crc kubenswrapper[4559]: I1123 06:58:40.548428 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 23 06:58:40 crc kubenswrapper[4559]: I1123 06:58:40.715782 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/7783a8fe-c77b-4e05-8c52-3e11d3a92196-cache\") pod \"swift-storage-0\" (UID: \"7783a8fe-c77b-4e05-8c52-3e11d3a92196\") " pod="openstack/swift-storage-0" Nov 23 06:58:40 crc kubenswrapper[4559]: I1123 06:58:40.715827 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/7783a8fe-c77b-4e05-8c52-3e11d3a92196-lock\") pod \"swift-storage-0\" (UID: \"7783a8fe-c77b-4e05-8c52-3e11d3a92196\") " pod="openstack/swift-storage-0" Nov 23 06:58:40 crc kubenswrapper[4559]: I1123 06:58:40.715851 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"7783a8fe-c77b-4e05-8c52-3e11d3a92196\") " pod="openstack/swift-storage-0" Nov 23 06:58:40 crc kubenswrapper[4559]: I1123 06:58:40.715870 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmpnk\" (UniqueName: \"kubernetes.io/projected/7783a8fe-c77b-4e05-8c52-3e11d3a92196-kube-api-access-fmpnk\") pod \"swift-storage-0\" (UID: \"7783a8fe-c77b-4e05-8c52-3e11d3a92196\") " pod="openstack/swift-storage-0" Nov 23 06:58:40 crc kubenswrapper[4559]: I1123 06:58:40.716008 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7783a8fe-c77b-4e05-8c52-3e11d3a92196-etc-swift\") pod \"swift-storage-0\" (UID: \"7783a8fe-c77b-4e05-8c52-3e11d3a92196\") " pod="openstack/swift-storage-0" Nov 23 06:58:40 crc kubenswrapper[4559]: I1123 06:58:40.818909 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/7783a8fe-c77b-4e05-8c52-3e11d3a92196-cache\") pod \"swift-storage-0\" (UID: \"7783a8fe-c77b-4e05-8c52-3e11d3a92196\") " pod="openstack/swift-storage-0" Nov 23 06:58:40 crc kubenswrapper[4559]: I1123 06:58:40.818975 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/7783a8fe-c77b-4e05-8c52-3e11d3a92196-lock\") pod \"swift-storage-0\" (UID: \"7783a8fe-c77b-4e05-8c52-3e11d3a92196\") " pod="openstack/swift-storage-0" Nov 23 06:58:40 crc kubenswrapper[4559]: I1123 06:58:40.819008 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"7783a8fe-c77b-4e05-8c52-3e11d3a92196\") " pod="openstack/swift-storage-0" Nov 23 06:58:40 crc kubenswrapper[4559]: I1123 06:58:40.819025 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmpnk\" (UniqueName: \"kubernetes.io/projected/7783a8fe-c77b-4e05-8c52-3e11d3a92196-kube-api-access-fmpnk\") pod \"swift-storage-0\" (UID: \"7783a8fe-c77b-4e05-8c52-3e11d3a92196\") " pod="openstack/swift-storage-0" Nov 23 06:58:40 crc kubenswrapper[4559]: I1123 06:58:40.819109 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7783a8fe-c77b-4e05-8c52-3e11d3a92196-etc-swift\") pod \"swift-storage-0\" (UID: \"7783a8fe-c77b-4e05-8c52-3e11d3a92196\") " pod="openstack/swift-storage-0" Nov 23 06:58:40 crc kubenswrapper[4559]: E1123 06:58:40.819390 4559 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 23 06:58:40 crc kubenswrapper[4559]: E1123 06:58:40.819419 4559 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 23 06:58:40 crc kubenswrapper[4559]: I1123 06:58:40.819430 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/7783a8fe-c77b-4e05-8c52-3e11d3a92196-cache\") pod \"swift-storage-0\" (UID: \"7783a8fe-c77b-4e05-8c52-3e11d3a92196\") " pod="openstack/swift-storage-0" Nov 23 06:58:40 crc kubenswrapper[4559]: I1123 06:58:40.819413 4559 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"7783a8fe-c77b-4e05-8c52-3e11d3a92196\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/swift-storage-0" Nov 23 06:58:40 crc kubenswrapper[4559]: E1123 06:58:40.819476 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7783a8fe-c77b-4e05-8c52-3e11d3a92196-etc-swift podName:7783a8fe-c77b-4e05-8c52-3e11d3a92196 nodeName:}" failed. No retries permitted until 2025-11-23 06:58:41.319458293 +0000 UTC m=+823.341443907 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7783a8fe-c77b-4e05-8c52-3e11d3a92196-etc-swift") pod "swift-storage-0" (UID: "7783a8fe-c77b-4e05-8c52-3e11d3a92196") : configmap "swift-ring-files" not found Nov 23 06:58:40 crc kubenswrapper[4559]: I1123 06:58:40.819476 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/7783a8fe-c77b-4e05-8c52-3e11d3a92196-lock\") pod \"swift-storage-0\" (UID: \"7783a8fe-c77b-4e05-8c52-3e11d3a92196\") " pod="openstack/swift-storage-0" Nov 23 06:58:40 crc kubenswrapper[4559]: I1123 06:58:40.840027 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmpnk\" (UniqueName: \"kubernetes.io/projected/7783a8fe-c77b-4e05-8c52-3e11d3a92196-kube-api-access-fmpnk\") pod \"swift-storage-0\" (UID: \"7783a8fe-c77b-4e05-8c52-3e11d3a92196\") " pod="openstack/swift-storage-0" Nov 23 06:58:40 crc kubenswrapper[4559]: I1123 06:58:40.849254 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"7783a8fe-c77b-4e05-8c52-3e11d3a92196\") " pod="openstack/swift-storage-0" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.036996 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-pzxk9"] Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.038628 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-pzxk9" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.042045 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.042583 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.058880 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.060955 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-pzxk9"] Nov 23 06:58:41 crc kubenswrapper[4559]: E1123 06:58:41.061525 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-hzb5h ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-hzb5h ring-data-devices scripts swiftconf]: context canceled" pod="openstack/swift-ring-rebalance-pzxk9" podUID="d4f143be-7318-4fca-9036-6a2a1c736fbe" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.089135 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-94gfd"] Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.090298 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-94gfd" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.093701 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-94gfd"] Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.097848 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-pzxk9"] Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.235270 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3696cad8-c9be-4efd-982a-a4c1d6de858b-combined-ca-bundle\") pod \"swift-ring-rebalance-94gfd\" (UID: \"3696cad8-c9be-4efd-982a-a4c1d6de858b\") " pod="openstack/swift-ring-rebalance-94gfd" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.235673 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lnnc\" (UniqueName: \"kubernetes.io/projected/3696cad8-c9be-4efd-982a-a4c1d6de858b-kube-api-access-5lnnc\") pod \"swift-ring-rebalance-94gfd\" (UID: \"3696cad8-c9be-4efd-982a-a4c1d6de858b\") " pod="openstack/swift-ring-rebalance-94gfd" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.235718 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3696cad8-c9be-4efd-982a-a4c1d6de858b-etc-swift\") pod \"swift-ring-rebalance-94gfd\" (UID: \"3696cad8-c9be-4efd-982a-a4c1d6de858b\") " pod="openstack/swift-ring-rebalance-94gfd" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.235741 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d4f143be-7318-4fca-9036-6a2a1c736fbe-scripts\") pod \"swift-ring-rebalance-pzxk9\" (UID: \"d4f143be-7318-4fca-9036-6a2a1c736fbe\") " pod="openstack/swift-ring-rebalance-pzxk9" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.236152 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3696cad8-c9be-4efd-982a-a4c1d6de858b-swiftconf\") pod \"swift-ring-rebalance-94gfd\" (UID: \"3696cad8-c9be-4efd-982a-a4c1d6de858b\") " pod="openstack/swift-ring-rebalance-94gfd" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.236198 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3696cad8-c9be-4efd-982a-a4c1d6de858b-ring-data-devices\") pod \"swift-ring-rebalance-94gfd\" (UID: \"3696cad8-c9be-4efd-982a-a4c1d6de858b\") " pod="openstack/swift-ring-rebalance-94gfd" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.236354 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3696cad8-c9be-4efd-982a-a4c1d6de858b-scripts\") pod \"swift-ring-rebalance-94gfd\" (UID: \"3696cad8-c9be-4efd-982a-a4c1d6de858b\") " pod="openstack/swift-ring-rebalance-94gfd" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.236433 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d4f143be-7318-4fca-9036-6a2a1c736fbe-ring-data-devices\") pod \"swift-ring-rebalance-pzxk9\" (UID: \"d4f143be-7318-4fca-9036-6a2a1c736fbe\") " pod="openstack/swift-ring-rebalance-pzxk9" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.236477 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzb5h\" (UniqueName: \"kubernetes.io/projected/d4f143be-7318-4fca-9036-6a2a1c736fbe-kube-api-access-hzb5h\") pod \"swift-ring-rebalance-pzxk9\" (UID: \"d4f143be-7318-4fca-9036-6a2a1c736fbe\") " pod="openstack/swift-ring-rebalance-pzxk9" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.236517 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d4f143be-7318-4fca-9036-6a2a1c736fbe-etc-swift\") pod \"swift-ring-rebalance-pzxk9\" (UID: \"d4f143be-7318-4fca-9036-6a2a1c736fbe\") " pod="openstack/swift-ring-rebalance-pzxk9" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.236595 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4f143be-7318-4fca-9036-6a2a1c736fbe-combined-ca-bundle\") pod \"swift-ring-rebalance-pzxk9\" (UID: \"d4f143be-7318-4fca-9036-6a2a1c736fbe\") " pod="openstack/swift-ring-rebalance-pzxk9" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.236635 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d4f143be-7318-4fca-9036-6a2a1c736fbe-swiftconf\") pod \"swift-ring-rebalance-pzxk9\" (UID: \"d4f143be-7318-4fca-9036-6a2a1c736fbe\") " pod="openstack/swift-ring-rebalance-pzxk9" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.236677 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3696cad8-c9be-4efd-982a-a4c1d6de858b-dispersionconf\") pod \"swift-ring-rebalance-94gfd\" (UID: \"3696cad8-c9be-4efd-982a-a4c1d6de858b\") " pod="openstack/swift-ring-rebalance-94gfd" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.236716 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d4f143be-7318-4fca-9036-6a2a1c736fbe-dispersionconf\") pod \"swift-ring-rebalance-pzxk9\" (UID: \"d4f143be-7318-4fca-9036-6a2a1c736fbe\") " pod="openstack/swift-ring-rebalance-pzxk9" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.338193 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3696cad8-c9be-4efd-982a-a4c1d6de858b-combined-ca-bundle\") pod \"swift-ring-rebalance-94gfd\" (UID: \"3696cad8-c9be-4efd-982a-a4c1d6de858b\") " pod="openstack/swift-ring-rebalance-94gfd" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.338466 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lnnc\" (UniqueName: \"kubernetes.io/projected/3696cad8-c9be-4efd-982a-a4c1d6de858b-kube-api-access-5lnnc\") pod \"swift-ring-rebalance-94gfd\" (UID: \"3696cad8-c9be-4efd-982a-a4c1d6de858b\") " pod="openstack/swift-ring-rebalance-94gfd" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.338526 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3696cad8-c9be-4efd-982a-a4c1d6de858b-etc-swift\") pod \"swift-ring-rebalance-94gfd\" (UID: \"3696cad8-c9be-4efd-982a-a4c1d6de858b\") " pod="openstack/swift-ring-rebalance-94gfd" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.338557 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d4f143be-7318-4fca-9036-6a2a1c736fbe-scripts\") pod \"swift-ring-rebalance-pzxk9\" (UID: \"d4f143be-7318-4fca-9036-6a2a1c736fbe\") " pod="openstack/swift-ring-rebalance-pzxk9" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.338628 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3696cad8-c9be-4efd-982a-a4c1d6de858b-swiftconf\") pod \"swift-ring-rebalance-94gfd\" (UID: \"3696cad8-c9be-4efd-982a-a4c1d6de858b\") " pod="openstack/swift-ring-rebalance-94gfd" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.338668 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3696cad8-c9be-4efd-982a-a4c1d6de858b-ring-data-devices\") pod \"swift-ring-rebalance-94gfd\" (UID: \"3696cad8-c9be-4efd-982a-a4c1d6de858b\") " pod="openstack/swift-ring-rebalance-94gfd" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.338730 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7783a8fe-c77b-4e05-8c52-3e11d3a92196-etc-swift\") pod \"swift-storage-0\" (UID: \"7783a8fe-c77b-4e05-8c52-3e11d3a92196\") " pod="openstack/swift-storage-0" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.338754 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3696cad8-c9be-4efd-982a-a4c1d6de858b-scripts\") pod \"swift-ring-rebalance-94gfd\" (UID: \"3696cad8-c9be-4efd-982a-a4c1d6de858b\") " pod="openstack/swift-ring-rebalance-94gfd" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.338784 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d4f143be-7318-4fca-9036-6a2a1c736fbe-ring-data-devices\") pod \"swift-ring-rebalance-pzxk9\" (UID: \"d4f143be-7318-4fca-9036-6a2a1c736fbe\") " pod="openstack/swift-ring-rebalance-pzxk9" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.338803 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzb5h\" (UniqueName: \"kubernetes.io/projected/d4f143be-7318-4fca-9036-6a2a1c736fbe-kube-api-access-hzb5h\") pod \"swift-ring-rebalance-pzxk9\" (UID: \"d4f143be-7318-4fca-9036-6a2a1c736fbe\") " pod="openstack/swift-ring-rebalance-pzxk9" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.338832 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d4f143be-7318-4fca-9036-6a2a1c736fbe-etc-swift\") pod \"swift-ring-rebalance-pzxk9\" (UID: \"d4f143be-7318-4fca-9036-6a2a1c736fbe\") " pod="openstack/swift-ring-rebalance-pzxk9" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.338896 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4f143be-7318-4fca-9036-6a2a1c736fbe-combined-ca-bundle\") pod \"swift-ring-rebalance-pzxk9\" (UID: \"d4f143be-7318-4fca-9036-6a2a1c736fbe\") " pod="openstack/swift-ring-rebalance-pzxk9" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.338911 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d4f143be-7318-4fca-9036-6a2a1c736fbe-swiftconf\") pod \"swift-ring-rebalance-pzxk9\" (UID: \"d4f143be-7318-4fca-9036-6a2a1c736fbe\") " pod="openstack/swift-ring-rebalance-pzxk9" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.338924 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3696cad8-c9be-4efd-982a-a4c1d6de858b-dispersionconf\") pod \"swift-ring-rebalance-94gfd\" (UID: \"3696cad8-c9be-4efd-982a-a4c1d6de858b\") " pod="openstack/swift-ring-rebalance-94gfd" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.338954 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d4f143be-7318-4fca-9036-6a2a1c736fbe-dispersionconf\") pod \"swift-ring-rebalance-pzxk9\" (UID: \"d4f143be-7318-4fca-9036-6a2a1c736fbe\") " pod="openstack/swift-ring-rebalance-pzxk9" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.339045 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3696cad8-c9be-4efd-982a-a4c1d6de858b-etc-swift\") pod \"swift-ring-rebalance-94gfd\" (UID: \"3696cad8-c9be-4efd-982a-a4c1d6de858b\") " pod="openstack/swift-ring-rebalance-94gfd" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.339428 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3696cad8-c9be-4efd-982a-a4c1d6de858b-ring-data-devices\") pod \"swift-ring-rebalance-94gfd\" (UID: \"3696cad8-c9be-4efd-982a-a4c1d6de858b\") " pod="openstack/swift-ring-rebalance-94gfd" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.339616 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d4f143be-7318-4fca-9036-6a2a1c736fbe-scripts\") pod \"swift-ring-rebalance-pzxk9\" (UID: \"d4f143be-7318-4fca-9036-6a2a1c736fbe\") " pod="openstack/swift-ring-rebalance-pzxk9" Nov 23 06:58:41 crc kubenswrapper[4559]: E1123 06:58:41.340229 4559 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 23 06:58:41 crc kubenswrapper[4559]: E1123 06:58:41.340268 4559 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 23 06:58:41 crc kubenswrapper[4559]: E1123 06:58:41.340304 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7783a8fe-c77b-4e05-8c52-3e11d3a92196-etc-swift podName:7783a8fe-c77b-4e05-8c52-3e11d3a92196 nodeName:}" failed. No retries permitted until 2025-11-23 06:58:42.340289877 +0000 UTC m=+824.362275491 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7783a8fe-c77b-4e05-8c52-3e11d3a92196-etc-swift") pod "swift-storage-0" (UID: "7783a8fe-c77b-4e05-8c52-3e11d3a92196") : configmap "swift-ring-files" not found Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.340354 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d4f143be-7318-4fca-9036-6a2a1c736fbe-etc-swift\") pod \"swift-ring-rebalance-pzxk9\" (UID: \"d4f143be-7318-4fca-9036-6a2a1c736fbe\") " pod="openstack/swift-ring-rebalance-pzxk9" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.340786 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3696cad8-c9be-4efd-982a-a4c1d6de858b-scripts\") pod \"swift-ring-rebalance-94gfd\" (UID: \"3696cad8-c9be-4efd-982a-a4c1d6de858b\") " pod="openstack/swift-ring-rebalance-94gfd" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.341038 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d4f143be-7318-4fca-9036-6a2a1c736fbe-ring-data-devices\") pod \"swift-ring-rebalance-pzxk9\" (UID: \"d4f143be-7318-4fca-9036-6a2a1c736fbe\") " pod="openstack/swift-ring-rebalance-pzxk9" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.343091 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d4f143be-7318-4fca-9036-6a2a1c736fbe-dispersionconf\") pod \"swift-ring-rebalance-pzxk9\" (UID: \"d4f143be-7318-4fca-9036-6a2a1c736fbe\") " pod="openstack/swift-ring-rebalance-pzxk9" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.343270 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3696cad8-c9be-4efd-982a-a4c1d6de858b-combined-ca-bundle\") pod \"swift-ring-rebalance-94gfd\" (UID: \"3696cad8-c9be-4efd-982a-a4c1d6de858b\") " pod="openstack/swift-ring-rebalance-94gfd" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.343948 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3696cad8-c9be-4efd-982a-a4c1d6de858b-dispersionconf\") pod \"swift-ring-rebalance-94gfd\" (UID: \"3696cad8-c9be-4efd-982a-a4c1d6de858b\") " pod="openstack/swift-ring-rebalance-94gfd" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.345087 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4f143be-7318-4fca-9036-6a2a1c736fbe-combined-ca-bundle\") pod \"swift-ring-rebalance-pzxk9\" (UID: \"d4f143be-7318-4fca-9036-6a2a1c736fbe\") " pod="openstack/swift-ring-rebalance-pzxk9" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.356878 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzb5h\" (UniqueName: \"kubernetes.io/projected/d4f143be-7318-4fca-9036-6a2a1c736fbe-kube-api-access-hzb5h\") pod \"swift-ring-rebalance-pzxk9\" (UID: \"d4f143be-7318-4fca-9036-6a2a1c736fbe\") " pod="openstack/swift-ring-rebalance-pzxk9" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.357033 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lnnc\" (UniqueName: \"kubernetes.io/projected/3696cad8-c9be-4efd-982a-a4c1d6de858b-kube-api-access-5lnnc\") pod \"swift-ring-rebalance-94gfd\" (UID: \"3696cad8-c9be-4efd-982a-a4c1d6de858b\") " pod="openstack/swift-ring-rebalance-94gfd" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.357516 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d4f143be-7318-4fca-9036-6a2a1c736fbe-swiftconf\") pod \"swift-ring-rebalance-pzxk9\" (UID: \"d4f143be-7318-4fca-9036-6a2a1c736fbe\") " pod="openstack/swift-ring-rebalance-pzxk9" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.362620 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3696cad8-c9be-4efd-982a-a4c1d6de858b-swiftconf\") pod \"swift-ring-rebalance-94gfd\" (UID: \"3696cad8-c9be-4efd-982a-a4c1d6de858b\") " pod="openstack/swift-ring-rebalance-94gfd" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.384486 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-pzxk9" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.385202 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7dfd8c6765-hrs74" event={"ID":"35d05808-5864-4a8f-ab67-e4627d8c8770","Type":"ContainerStarted","Data":"991c1fdd9ab230c51a73e633f16b0abb7617d383c7887c06e344965a25700f6a"} Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.385605 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7dfd8c6765-hrs74" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.403517 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-94gfd" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.408663 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7dfd8c6765-hrs74" podStartSLOduration=2.4086349240000002 podStartE2EDuration="2.408634924s" podCreationTimestamp="2025-11-23 06:58:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:58:41.404258825 +0000 UTC m=+823.426244439" watchObservedRunningTime="2025-11-23 06:58:41.408634924 +0000 UTC m=+823.430620538" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.419110 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-pzxk9" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.541421 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzb5h\" (UniqueName: \"kubernetes.io/projected/d4f143be-7318-4fca-9036-6a2a1c736fbe-kube-api-access-hzb5h\") pod \"d4f143be-7318-4fca-9036-6a2a1c736fbe\" (UID: \"d4f143be-7318-4fca-9036-6a2a1c736fbe\") " Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.541462 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d4f143be-7318-4fca-9036-6a2a1c736fbe-swiftconf\") pod \"d4f143be-7318-4fca-9036-6a2a1c736fbe\" (UID: \"d4f143be-7318-4fca-9036-6a2a1c736fbe\") " Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.541515 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4f143be-7318-4fca-9036-6a2a1c736fbe-combined-ca-bundle\") pod \"d4f143be-7318-4fca-9036-6a2a1c736fbe\" (UID: \"d4f143be-7318-4fca-9036-6a2a1c736fbe\") " Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.541700 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d4f143be-7318-4fca-9036-6a2a1c736fbe-ring-data-devices\") pod \"d4f143be-7318-4fca-9036-6a2a1c736fbe\" (UID: \"d4f143be-7318-4fca-9036-6a2a1c736fbe\") " Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.541818 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d4f143be-7318-4fca-9036-6a2a1c736fbe-scripts\") pod \"d4f143be-7318-4fca-9036-6a2a1c736fbe\" (UID: \"d4f143be-7318-4fca-9036-6a2a1c736fbe\") " Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.541858 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d4f143be-7318-4fca-9036-6a2a1c736fbe-etc-swift\") pod \"d4f143be-7318-4fca-9036-6a2a1c736fbe\" (UID: \"d4f143be-7318-4fca-9036-6a2a1c736fbe\") " Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.541891 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d4f143be-7318-4fca-9036-6a2a1c736fbe-dispersionconf\") pod \"d4f143be-7318-4fca-9036-6a2a1c736fbe\" (UID: \"d4f143be-7318-4fca-9036-6a2a1c736fbe\") " Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.543101 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4f143be-7318-4fca-9036-6a2a1c736fbe-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "d4f143be-7318-4fca-9036-6a2a1c736fbe" (UID: "d4f143be-7318-4fca-9036-6a2a1c736fbe"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.543212 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4f143be-7318-4fca-9036-6a2a1c736fbe-scripts" (OuterVolumeSpecName: "scripts") pod "d4f143be-7318-4fca-9036-6a2a1c736fbe" (UID: "d4f143be-7318-4fca-9036-6a2a1c736fbe"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.543388 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4f143be-7318-4fca-9036-6a2a1c736fbe-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "d4f143be-7318-4fca-9036-6a2a1c736fbe" (UID: "d4f143be-7318-4fca-9036-6a2a1c736fbe"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.545358 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4f143be-7318-4fca-9036-6a2a1c736fbe-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "d4f143be-7318-4fca-9036-6a2a1c736fbe" (UID: "d4f143be-7318-4fca-9036-6a2a1c736fbe"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.546242 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4f143be-7318-4fca-9036-6a2a1c736fbe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4f143be-7318-4fca-9036-6a2a1c736fbe" (UID: "d4f143be-7318-4fca-9036-6a2a1c736fbe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.546544 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4f143be-7318-4fca-9036-6a2a1c736fbe-kube-api-access-hzb5h" (OuterVolumeSpecName: "kube-api-access-hzb5h") pod "d4f143be-7318-4fca-9036-6a2a1c736fbe" (UID: "d4f143be-7318-4fca-9036-6a2a1c736fbe"). InnerVolumeSpecName "kube-api-access-hzb5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.547835 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4f143be-7318-4fca-9036-6a2a1c736fbe-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "d4f143be-7318-4fca-9036-6a2a1c736fbe" (UID: "d4f143be-7318-4fca-9036-6a2a1c736fbe"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.643299 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4f143be-7318-4fca-9036-6a2a1c736fbe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.643504 4559 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d4f143be-7318-4fca-9036-6a2a1c736fbe-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.643514 4559 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d4f143be-7318-4fca-9036-6a2a1c736fbe-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.643523 4559 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d4f143be-7318-4fca-9036-6a2a1c736fbe-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.643530 4559 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d4f143be-7318-4fca-9036-6a2a1c736fbe-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.643538 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzb5h\" (UniqueName: \"kubernetes.io/projected/d4f143be-7318-4fca-9036-6a2a1c736fbe-kube-api-access-hzb5h\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.643547 4559 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d4f143be-7318-4fca-9036-6a2a1c736fbe-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:41 crc kubenswrapper[4559]: I1123 06:58:41.790832 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-94gfd"] Nov 23 06:58:41 crc kubenswrapper[4559]: W1123 06:58:41.802309 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3696cad8_c9be_4efd_982a_a4c1d6de858b.slice/crio-b919558b67bf3e2ddc28345c59d249643aab8dfa5bcabcfff0dedaa84cb15193 WatchSource:0}: Error finding container b919558b67bf3e2ddc28345c59d249643aab8dfa5bcabcfff0dedaa84cb15193: Status 404 returned error can't find the container with id b919558b67bf3e2ddc28345c59d249643aab8dfa5bcabcfff0dedaa84cb15193 Nov 23 06:58:42 crc kubenswrapper[4559]: I1123 06:58:42.354861 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7783a8fe-c77b-4e05-8c52-3e11d3a92196-etc-swift\") pod \"swift-storage-0\" (UID: \"7783a8fe-c77b-4e05-8c52-3e11d3a92196\") " pod="openstack/swift-storage-0" Nov 23 06:58:42 crc kubenswrapper[4559]: E1123 06:58:42.355030 4559 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 23 06:58:42 crc kubenswrapper[4559]: E1123 06:58:42.355142 4559 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 23 06:58:42 crc kubenswrapper[4559]: E1123 06:58:42.355201 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7783a8fe-c77b-4e05-8c52-3e11d3a92196-etc-swift podName:7783a8fe-c77b-4e05-8c52-3e11d3a92196 nodeName:}" failed. No retries permitted until 2025-11-23 06:58:44.355182295 +0000 UTC m=+826.377167909 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7783a8fe-c77b-4e05-8c52-3e11d3a92196-etc-swift") pod "swift-storage-0" (UID: "7783a8fe-c77b-4e05-8c52-3e11d3a92196") : configmap "swift-ring-files" not found Nov 23 06:58:42 crc kubenswrapper[4559]: I1123 06:58:42.393463 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-pzxk9" Nov 23 06:58:42 crc kubenswrapper[4559]: I1123 06:58:42.393946 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-94gfd" event={"ID":"3696cad8-c9be-4efd-982a-a4c1d6de858b","Type":"ContainerStarted","Data":"b919558b67bf3e2ddc28345c59d249643aab8dfa5bcabcfff0dedaa84cb15193"} Nov 23 06:58:42 crc kubenswrapper[4559]: I1123 06:58:42.440981 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-pzxk9"] Nov 23 06:58:42 crc kubenswrapper[4559]: I1123 06:58:42.444774 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-pzxk9"] Nov 23 06:58:44 crc kubenswrapper[4559]: I1123 06:58:44.285111 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4f143be-7318-4fca-9036-6a2a1c736fbe" path="/var/lib/kubelet/pods/d4f143be-7318-4fca-9036-6a2a1c736fbe/volumes" Nov 23 06:58:44 crc kubenswrapper[4559]: I1123 06:58:44.384725 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7783a8fe-c77b-4e05-8c52-3e11d3a92196-etc-swift\") pod \"swift-storage-0\" (UID: \"7783a8fe-c77b-4e05-8c52-3e11d3a92196\") " pod="openstack/swift-storage-0" Nov 23 06:58:44 crc kubenswrapper[4559]: E1123 06:58:44.384944 4559 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 23 06:58:44 crc kubenswrapper[4559]: E1123 06:58:44.384981 4559 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 23 06:58:44 crc kubenswrapper[4559]: E1123 06:58:44.385046 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7783a8fe-c77b-4e05-8c52-3e11d3a92196-etc-swift podName:7783a8fe-c77b-4e05-8c52-3e11d3a92196 nodeName:}" failed. No retries permitted until 2025-11-23 06:58:48.385026311 +0000 UTC m=+830.407011925 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7783a8fe-c77b-4e05-8c52-3e11d3a92196-etc-swift") pod "swift-storage-0" (UID: "7783a8fe-c77b-4e05-8c52-3e11d3a92196") : configmap "swift-ring-files" not found Nov 23 06:58:46 crc kubenswrapper[4559]: I1123 06:58:46.429669 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-94gfd" event={"ID":"3696cad8-c9be-4efd-982a-a4c1d6de858b","Type":"ContainerStarted","Data":"4bdc3fc819475e4cd327eda031fc6edb68c9c8a54b713cfdf8ed3e33b4e26531"} Nov 23 06:58:46 crc kubenswrapper[4559]: I1123 06:58:46.453773 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-94gfd" podStartSLOduration=1.269222224 podStartE2EDuration="5.453756187s" podCreationTimestamp="2025-11-23 06:58:41 +0000 UTC" firstStartedPulling="2025-11-23 06:58:41.804929708 +0000 UTC m=+823.826915321" lastFinishedPulling="2025-11-23 06:58:45.98946367 +0000 UTC m=+828.011449284" observedRunningTime="2025-11-23 06:58:46.448493481 +0000 UTC m=+828.470479095" watchObservedRunningTime="2025-11-23 06:58:46.453756187 +0000 UTC m=+828.475741802" Nov 23 06:58:48 crc kubenswrapper[4559]: I1123 06:58:48.448211 4559 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-blj85" podUID="20662e83-ba79-4c97-80fb-98fda28c1149" containerName="ovn-controller" probeResult="failure" output=< Nov 23 06:58:48 crc kubenswrapper[4559]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 23 06:58:48 crc kubenswrapper[4559]: > Nov 23 06:58:48 crc kubenswrapper[4559]: I1123 06:58:48.454464 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7783a8fe-c77b-4e05-8c52-3e11d3a92196-etc-swift\") pod \"swift-storage-0\" (UID: \"7783a8fe-c77b-4e05-8c52-3e11d3a92196\") " pod="openstack/swift-storage-0" Nov 23 06:58:48 crc kubenswrapper[4559]: E1123 06:58:48.454683 4559 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 23 06:58:48 crc kubenswrapper[4559]: E1123 06:58:48.454699 4559 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 23 06:58:48 crc kubenswrapper[4559]: E1123 06:58:48.454765 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7783a8fe-c77b-4e05-8c52-3e11d3a92196-etc-swift podName:7783a8fe-c77b-4e05-8c52-3e11d3a92196 nodeName:}" failed. No retries permitted until 2025-11-23 06:58:56.454745415 +0000 UTC m=+838.476731029 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7783a8fe-c77b-4e05-8c52-3e11d3a92196-etc-swift") pod "swift-storage-0" (UID: "7783a8fe-c77b-4e05-8c52-3e11d3a92196") : configmap "swift-ring-files" not found Nov 23 06:58:48 crc kubenswrapper[4559]: I1123 06:58:48.458113 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-jg7hk" Nov 23 06:58:48 crc kubenswrapper[4559]: I1123 06:58:48.473146 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-jg7hk" Nov 23 06:58:48 crc kubenswrapper[4559]: I1123 06:58:48.678251 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-blj85-config-mrnpd"] Nov 23 06:58:48 crc kubenswrapper[4559]: I1123 06:58:48.679623 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-blj85-config-mrnpd" Nov 23 06:58:48 crc kubenswrapper[4559]: I1123 06:58:48.681493 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 23 06:58:48 crc kubenswrapper[4559]: I1123 06:58:48.684356 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-blj85-config-mrnpd"] Nov 23 06:58:48 crc kubenswrapper[4559]: I1123 06:58:48.759333 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0e7f2ae1-1734-49cb-998e-efe380cc6c47-var-run\") pod \"ovn-controller-blj85-config-mrnpd\" (UID: \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\") " pod="openstack/ovn-controller-blj85-config-mrnpd" Nov 23 06:58:48 crc kubenswrapper[4559]: I1123 06:58:48.759378 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0e7f2ae1-1734-49cb-998e-efe380cc6c47-scripts\") pod \"ovn-controller-blj85-config-mrnpd\" (UID: \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\") " pod="openstack/ovn-controller-blj85-config-mrnpd" Nov 23 06:58:48 crc kubenswrapper[4559]: I1123 06:58:48.759424 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46fjt\" (UniqueName: \"kubernetes.io/projected/0e7f2ae1-1734-49cb-998e-efe380cc6c47-kube-api-access-46fjt\") pod \"ovn-controller-blj85-config-mrnpd\" (UID: \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\") " pod="openstack/ovn-controller-blj85-config-mrnpd" Nov 23 06:58:48 crc kubenswrapper[4559]: I1123 06:58:48.759584 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0e7f2ae1-1734-49cb-998e-efe380cc6c47-var-run-ovn\") pod \"ovn-controller-blj85-config-mrnpd\" (UID: \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\") " pod="openstack/ovn-controller-blj85-config-mrnpd" Nov 23 06:58:48 crc kubenswrapper[4559]: I1123 06:58:48.759703 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0e7f2ae1-1734-49cb-998e-efe380cc6c47-var-log-ovn\") pod \"ovn-controller-blj85-config-mrnpd\" (UID: \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\") " pod="openstack/ovn-controller-blj85-config-mrnpd" Nov 23 06:58:48 crc kubenswrapper[4559]: I1123 06:58:48.759789 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0e7f2ae1-1734-49cb-998e-efe380cc6c47-additional-scripts\") pod \"ovn-controller-blj85-config-mrnpd\" (UID: \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\") " pod="openstack/ovn-controller-blj85-config-mrnpd" Nov 23 06:58:48 crc kubenswrapper[4559]: I1123 06:58:48.861589 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0e7f2ae1-1734-49cb-998e-efe380cc6c47-additional-scripts\") pod \"ovn-controller-blj85-config-mrnpd\" (UID: \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\") " pod="openstack/ovn-controller-blj85-config-mrnpd" Nov 23 06:58:48 crc kubenswrapper[4559]: I1123 06:58:48.861669 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0e7f2ae1-1734-49cb-998e-efe380cc6c47-var-run\") pod \"ovn-controller-blj85-config-mrnpd\" (UID: \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\") " pod="openstack/ovn-controller-blj85-config-mrnpd" Nov 23 06:58:48 crc kubenswrapper[4559]: I1123 06:58:48.861701 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0e7f2ae1-1734-49cb-998e-efe380cc6c47-scripts\") pod \"ovn-controller-blj85-config-mrnpd\" (UID: \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\") " pod="openstack/ovn-controller-blj85-config-mrnpd" Nov 23 06:58:48 crc kubenswrapper[4559]: I1123 06:58:48.861745 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46fjt\" (UniqueName: \"kubernetes.io/projected/0e7f2ae1-1734-49cb-998e-efe380cc6c47-kube-api-access-46fjt\") pod \"ovn-controller-blj85-config-mrnpd\" (UID: \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\") " pod="openstack/ovn-controller-blj85-config-mrnpd" Nov 23 06:58:48 crc kubenswrapper[4559]: I1123 06:58:48.861793 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0e7f2ae1-1734-49cb-998e-efe380cc6c47-var-run-ovn\") pod \"ovn-controller-blj85-config-mrnpd\" (UID: \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\") " pod="openstack/ovn-controller-blj85-config-mrnpd" Nov 23 06:58:48 crc kubenswrapper[4559]: I1123 06:58:48.861836 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0e7f2ae1-1734-49cb-998e-efe380cc6c47-var-log-ovn\") pod \"ovn-controller-blj85-config-mrnpd\" (UID: \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\") " pod="openstack/ovn-controller-blj85-config-mrnpd" Nov 23 06:58:48 crc kubenswrapper[4559]: I1123 06:58:48.862122 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0e7f2ae1-1734-49cb-998e-efe380cc6c47-var-log-ovn\") pod \"ovn-controller-blj85-config-mrnpd\" (UID: \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\") " pod="openstack/ovn-controller-blj85-config-mrnpd" Nov 23 06:58:48 crc kubenswrapper[4559]: I1123 06:58:48.862115 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0e7f2ae1-1734-49cb-998e-efe380cc6c47-var-run\") pod \"ovn-controller-blj85-config-mrnpd\" (UID: \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\") " pod="openstack/ovn-controller-blj85-config-mrnpd" Nov 23 06:58:48 crc kubenswrapper[4559]: I1123 06:58:48.862561 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0e7f2ae1-1734-49cb-998e-efe380cc6c47-var-run-ovn\") pod \"ovn-controller-blj85-config-mrnpd\" (UID: \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\") " pod="openstack/ovn-controller-blj85-config-mrnpd" Nov 23 06:58:48 crc kubenswrapper[4559]: I1123 06:58:48.864048 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0e7f2ae1-1734-49cb-998e-efe380cc6c47-additional-scripts\") pod \"ovn-controller-blj85-config-mrnpd\" (UID: \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\") " pod="openstack/ovn-controller-blj85-config-mrnpd" Nov 23 06:58:48 crc kubenswrapper[4559]: I1123 06:58:48.864703 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0e7f2ae1-1734-49cb-998e-efe380cc6c47-scripts\") pod \"ovn-controller-blj85-config-mrnpd\" (UID: \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\") " pod="openstack/ovn-controller-blj85-config-mrnpd" Nov 23 06:58:48 crc kubenswrapper[4559]: I1123 06:58:48.880120 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46fjt\" (UniqueName: \"kubernetes.io/projected/0e7f2ae1-1734-49cb-998e-efe380cc6c47-kube-api-access-46fjt\") pod \"ovn-controller-blj85-config-mrnpd\" (UID: \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\") " pod="openstack/ovn-controller-blj85-config-mrnpd" Nov 23 06:58:49 crc kubenswrapper[4559]: I1123 06:58:49.005829 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-blj85-config-mrnpd" Nov 23 06:58:49 crc kubenswrapper[4559]: I1123 06:58:49.437002 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-blj85-config-mrnpd"] Nov 23 06:58:49 crc kubenswrapper[4559]: I1123 06:58:49.455102 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-blj85-config-mrnpd" event={"ID":"0e7f2ae1-1734-49cb-998e-efe380cc6c47","Type":"ContainerStarted","Data":"b164ca75df46140716aa0764387fb5fb843d2da2c97a5864be5e853f74cfd5f5"} Nov 23 06:58:49 crc kubenswrapper[4559]: I1123 06:58:49.749789 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7dfd8c6765-hrs74" Nov 23 06:58:49 crc kubenswrapper[4559]: I1123 06:58:49.826581 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b8455895f-w72ch"] Nov 23 06:58:49 crc kubenswrapper[4559]: I1123 06:58:49.826932 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b8455895f-w72ch" podUID="4460f104-a577-49e6-bb5a-dc50cb3401eb" containerName="dnsmasq-dns" containerID="cri-o://a0141bb4b5a5004234d5f211f1fb42f3f155c25253aa7cce564a0e938e8bc619" gracePeriod=10 Nov 23 06:58:50 crc kubenswrapper[4559]: I1123 06:58:50.468561 4559 generic.go:334] "Generic (PLEG): container finished" podID="4460f104-a577-49e6-bb5a-dc50cb3401eb" containerID="a0141bb4b5a5004234d5f211f1fb42f3f155c25253aa7cce564a0e938e8bc619" exitCode=0 Nov 23 06:58:50 crc kubenswrapper[4559]: I1123 06:58:50.468664 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b8455895f-w72ch" event={"ID":"4460f104-a577-49e6-bb5a-dc50cb3401eb","Type":"ContainerDied","Data":"a0141bb4b5a5004234d5f211f1fb42f3f155c25253aa7cce564a0e938e8bc619"} Nov 23 06:58:50 crc kubenswrapper[4559]: I1123 06:58:50.470608 4559 generic.go:334] "Generic (PLEG): container finished" podID="0e7f2ae1-1734-49cb-998e-efe380cc6c47" containerID="a5990d9b69ba08df670ad9a966377f748277a1de4d043ccfe6259d49aa74c25c" exitCode=0 Nov 23 06:58:50 crc kubenswrapper[4559]: I1123 06:58:50.470672 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-blj85-config-mrnpd" event={"ID":"0e7f2ae1-1734-49cb-998e-efe380cc6c47","Type":"ContainerDied","Data":"a5990d9b69ba08df670ad9a966377f748277a1de4d043ccfe6259d49aa74c25c"} Nov 23 06:58:52 crc kubenswrapper[4559]: I1123 06:58:52.490168 4559 generic.go:334] "Generic (PLEG): container finished" podID="3696cad8-c9be-4efd-982a-a4c1d6de858b" containerID="4bdc3fc819475e4cd327eda031fc6edb68c9c8a54b713cfdf8ed3e33b4e26531" exitCode=0 Nov 23 06:58:52 crc kubenswrapper[4559]: I1123 06:58:52.490247 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-94gfd" event={"ID":"3696cad8-c9be-4efd-982a-a4c1d6de858b","Type":"ContainerDied","Data":"4bdc3fc819475e4cd327eda031fc6edb68c9c8a54b713cfdf8ed3e33b4e26531"} Nov 23 06:58:52 crc kubenswrapper[4559]: I1123 06:58:52.990045 4559 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b8455895f-w72ch" podUID="4460f104-a577-49e6-bb5a-dc50cb3401eb" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.111:5353: connect: connection refused" Nov 23 06:58:53 crc kubenswrapper[4559]: I1123 06:58:53.445175 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-blj85" Nov 23 06:58:53 crc kubenswrapper[4559]: I1123 06:58:53.502410 4559 generic.go:334] "Generic (PLEG): container finished" podID="a0fbc3f6-1c87-495a-a38f-5a0ae768ee20" containerID="832a304e7da998cd94f31c322c8f84bedc01786f45ac24347d345e9f99aab1a6" exitCode=0 Nov 23 06:58:53 crc kubenswrapper[4559]: I1123 06:58:53.502509 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20","Type":"ContainerDied","Data":"832a304e7da998cd94f31c322c8f84bedc01786f45ac24347d345e9f99aab1a6"} Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.682690 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-94gfd" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.687110 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-blj85-config-mrnpd" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.716068 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3696cad8-c9be-4efd-982a-a4c1d6de858b-ring-data-devices\") pod \"3696cad8-c9be-4efd-982a-a4c1d6de858b\" (UID: \"3696cad8-c9be-4efd-982a-a4c1d6de858b\") " Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.716126 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3696cad8-c9be-4efd-982a-a4c1d6de858b-swiftconf\") pod \"3696cad8-c9be-4efd-982a-a4c1d6de858b\" (UID: \"3696cad8-c9be-4efd-982a-a4c1d6de858b\") " Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.716186 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0e7f2ae1-1734-49cb-998e-efe380cc6c47-additional-scripts\") pod \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\" (UID: \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\") " Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.716209 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0e7f2ae1-1734-49cb-998e-efe380cc6c47-var-log-ovn\") pod \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\" (UID: \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\") " Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.716262 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lnnc\" (UniqueName: \"kubernetes.io/projected/3696cad8-c9be-4efd-982a-a4c1d6de858b-kube-api-access-5lnnc\") pod \"3696cad8-c9be-4efd-982a-a4c1d6de858b\" (UID: \"3696cad8-c9be-4efd-982a-a4c1d6de858b\") " Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.716284 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3696cad8-c9be-4efd-982a-a4c1d6de858b-combined-ca-bundle\") pod \"3696cad8-c9be-4efd-982a-a4c1d6de858b\" (UID: \"3696cad8-c9be-4efd-982a-a4c1d6de858b\") " Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.716346 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3696cad8-c9be-4efd-982a-a4c1d6de858b-etc-swift\") pod \"3696cad8-c9be-4efd-982a-a4c1d6de858b\" (UID: \"3696cad8-c9be-4efd-982a-a4c1d6de858b\") " Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.716390 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0e7f2ae1-1734-49cb-998e-efe380cc6c47-scripts\") pod \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\" (UID: \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\") " Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.716426 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0e7f2ae1-1734-49cb-998e-efe380cc6c47-var-run-ovn\") pod \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\" (UID: \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\") " Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.716463 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3696cad8-c9be-4efd-982a-a4c1d6de858b-scripts\") pod \"3696cad8-c9be-4efd-982a-a4c1d6de858b\" (UID: \"3696cad8-c9be-4efd-982a-a4c1d6de858b\") " Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.716633 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46fjt\" (UniqueName: \"kubernetes.io/projected/0e7f2ae1-1734-49cb-998e-efe380cc6c47-kube-api-access-46fjt\") pod \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\" (UID: \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\") " Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.717708 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0e7f2ae1-1734-49cb-998e-efe380cc6c47-var-run\") pod \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\" (UID: \"0e7f2ae1-1734-49cb-998e-efe380cc6c47\") " Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.717753 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3696cad8-c9be-4efd-982a-a4c1d6de858b-dispersionconf\") pod \"3696cad8-c9be-4efd-982a-a4c1d6de858b\" (UID: \"3696cad8-c9be-4efd-982a-a4c1d6de858b\") " Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.718503 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0e7f2ae1-1734-49cb-998e-efe380cc6c47-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "0e7f2ae1-1734-49cb-998e-efe380cc6c47" (UID: "0e7f2ae1-1734-49cb-998e-efe380cc6c47"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.719142 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0e7f2ae1-1734-49cb-998e-efe380cc6c47-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "0e7f2ae1-1734-49cb-998e-efe380cc6c47" (UID: "0e7f2ae1-1734-49cb-998e-efe380cc6c47"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.719280 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0e7f2ae1-1734-49cb-998e-efe380cc6c47-var-run" (OuterVolumeSpecName: "var-run") pod "0e7f2ae1-1734-49cb-998e-efe380cc6c47" (UID: "0e7f2ae1-1734-49cb-998e-efe380cc6c47"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.719434 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e7f2ae1-1734-49cb-998e-efe380cc6c47-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "0e7f2ae1-1734-49cb-998e-efe380cc6c47" (UID: "0e7f2ae1-1734-49cb-998e-efe380cc6c47"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.719444 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3696cad8-c9be-4efd-982a-a4c1d6de858b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "3696cad8-c9be-4efd-982a-a4c1d6de858b" (UID: "3696cad8-c9be-4efd-982a-a4c1d6de858b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.719629 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3696cad8-c9be-4efd-982a-a4c1d6de858b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "3696cad8-c9be-4efd-982a-a4c1d6de858b" (UID: "3696cad8-c9be-4efd-982a-a4c1d6de858b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.721987 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e7f2ae1-1734-49cb-998e-efe380cc6c47-scripts" (OuterVolumeSpecName: "scripts") pod "0e7f2ae1-1734-49cb-998e-efe380cc6c47" (UID: "0e7f2ae1-1734-49cb-998e-efe380cc6c47"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.737967 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e7f2ae1-1734-49cb-998e-efe380cc6c47-kube-api-access-46fjt" (OuterVolumeSpecName: "kube-api-access-46fjt") pod "0e7f2ae1-1734-49cb-998e-efe380cc6c47" (UID: "0e7f2ae1-1734-49cb-998e-efe380cc6c47"). InnerVolumeSpecName "kube-api-access-46fjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.738158 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3696cad8-c9be-4efd-982a-a4c1d6de858b-kube-api-access-5lnnc" (OuterVolumeSpecName: "kube-api-access-5lnnc") pod "3696cad8-c9be-4efd-982a-a4c1d6de858b" (UID: "3696cad8-c9be-4efd-982a-a4c1d6de858b"). InnerVolumeSpecName "kube-api-access-5lnnc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.738567 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3696cad8-c9be-4efd-982a-a4c1d6de858b-scripts" (OuterVolumeSpecName: "scripts") pod "3696cad8-c9be-4efd-982a-a4c1d6de858b" (UID: "3696cad8-c9be-4efd-982a-a4c1d6de858b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.740692 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3696cad8-c9be-4efd-982a-a4c1d6de858b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "3696cad8-c9be-4efd-982a-a4c1d6de858b" (UID: "3696cad8-c9be-4efd-982a-a4c1d6de858b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.746812 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3696cad8-c9be-4efd-982a-a4c1d6de858b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3696cad8-c9be-4efd-982a-a4c1d6de858b" (UID: "3696cad8-c9be-4efd-982a-a4c1d6de858b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.751562 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3696cad8-c9be-4efd-982a-a4c1d6de858b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "3696cad8-c9be-4efd-982a-a4c1d6de858b" (UID: "3696cad8-c9be-4efd-982a-a4c1d6de858b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.779598 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b8455895f-w72ch" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.819412 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4460f104-a577-49e6-bb5a-dc50cb3401eb-ovsdbserver-nb\") pod \"4460f104-a577-49e6-bb5a-dc50cb3401eb\" (UID: \"4460f104-a577-49e6-bb5a-dc50cb3401eb\") " Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.819705 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjckk\" (UniqueName: \"kubernetes.io/projected/4460f104-a577-49e6-bb5a-dc50cb3401eb-kube-api-access-vjckk\") pod \"4460f104-a577-49e6-bb5a-dc50cb3401eb\" (UID: \"4460f104-a577-49e6-bb5a-dc50cb3401eb\") " Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.819776 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4460f104-a577-49e6-bb5a-dc50cb3401eb-dns-svc\") pod \"4460f104-a577-49e6-bb5a-dc50cb3401eb\" (UID: \"4460f104-a577-49e6-bb5a-dc50cb3401eb\") " Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.819807 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4460f104-a577-49e6-bb5a-dc50cb3401eb-config\") pod \"4460f104-a577-49e6-bb5a-dc50cb3401eb\" (UID: \"4460f104-a577-49e6-bb5a-dc50cb3401eb\") " Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.820199 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4460f104-a577-49e6-bb5a-dc50cb3401eb-ovsdbserver-sb\") pod \"4460f104-a577-49e6-bb5a-dc50cb3401eb\" (UID: \"4460f104-a577-49e6-bb5a-dc50cb3401eb\") " Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.821203 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46fjt\" (UniqueName: \"kubernetes.io/projected/0e7f2ae1-1734-49cb-998e-efe380cc6c47-kube-api-access-46fjt\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.821224 4559 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0e7f2ae1-1734-49cb-998e-efe380cc6c47-var-run\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.821232 4559 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3696cad8-c9be-4efd-982a-a4c1d6de858b-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.821242 4559 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3696cad8-c9be-4efd-982a-a4c1d6de858b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.821250 4559 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3696cad8-c9be-4efd-982a-a4c1d6de858b-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.821258 4559 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0e7f2ae1-1734-49cb-998e-efe380cc6c47-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.821265 4559 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0e7f2ae1-1734-49cb-998e-efe380cc6c47-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.821273 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lnnc\" (UniqueName: \"kubernetes.io/projected/3696cad8-c9be-4efd-982a-a4c1d6de858b-kube-api-access-5lnnc\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.821280 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3696cad8-c9be-4efd-982a-a4c1d6de858b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.821288 4559 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3696cad8-c9be-4efd-982a-a4c1d6de858b-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.821298 4559 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0e7f2ae1-1734-49cb-998e-efe380cc6c47-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.821305 4559 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0e7f2ae1-1734-49cb-998e-efe380cc6c47-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.821313 4559 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3696cad8-c9be-4efd-982a-a4c1d6de858b-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.823428 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4460f104-a577-49e6-bb5a-dc50cb3401eb-kube-api-access-vjckk" (OuterVolumeSpecName: "kube-api-access-vjckk") pod "4460f104-a577-49e6-bb5a-dc50cb3401eb" (UID: "4460f104-a577-49e6-bb5a-dc50cb3401eb"). InnerVolumeSpecName "kube-api-access-vjckk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.852629 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4460f104-a577-49e6-bb5a-dc50cb3401eb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4460f104-a577-49e6-bb5a-dc50cb3401eb" (UID: "4460f104-a577-49e6-bb5a-dc50cb3401eb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.856237 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4460f104-a577-49e6-bb5a-dc50cb3401eb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4460f104-a577-49e6-bb5a-dc50cb3401eb" (UID: "4460f104-a577-49e6-bb5a-dc50cb3401eb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.857046 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4460f104-a577-49e6-bb5a-dc50cb3401eb-config" (OuterVolumeSpecName: "config") pod "4460f104-a577-49e6-bb5a-dc50cb3401eb" (UID: "4460f104-a577-49e6-bb5a-dc50cb3401eb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.861107 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4460f104-a577-49e6-bb5a-dc50cb3401eb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4460f104-a577-49e6-bb5a-dc50cb3401eb" (UID: "4460f104-a577-49e6-bb5a-dc50cb3401eb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.922559 4559 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4460f104-a577-49e6-bb5a-dc50cb3401eb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.922588 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjckk\" (UniqueName: \"kubernetes.io/projected/4460f104-a577-49e6-bb5a-dc50cb3401eb-kube-api-access-vjckk\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.922598 4559 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4460f104-a577-49e6-bb5a-dc50cb3401eb-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.922608 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4460f104-a577-49e6-bb5a-dc50cb3401eb-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:55 crc kubenswrapper[4559]: I1123 06:58:55.922617 4559 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4460f104-a577-49e6-bb5a-dc50cb3401eb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 06:58:56 crc kubenswrapper[4559]: I1123 06:58:56.529962 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-dwbhb" event={"ID":"a5cc366d-e466-476d-a318-1582a4de4a93","Type":"ContainerStarted","Data":"28c045ec04f1f1fbec6423600a427ac6ae17f3d9f67ebec610f72f1a248b0593"} Nov 23 06:58:56 crc kubenswrapper[4559]: I1123 06:58:56.531028 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7783a8fe-c77b-4e05-8c52-3e11d3a92196-etc-swift\") pod \"swift-storage-0\" (UID: \"7783a8fe-c77b-4e05-8c52-3e11d3a92196\") " pod="openstack/swift-storage-0" Nov 23 06:58:56 crc kubenswrapper[4559]: I1123 06:58:56.531813 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20","Type":"ContainerStarted","Data":"059e75026cf507f704dab0c20ddda3da72f5483549bf88ab566bab645ad7eb61"} Nov 23 06:58:56 crc kubenswrapper[4559]: I1123 06:58:56.531976 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 23 06:58:56 crc kubenswrapper[4559]: I1123 06:58:56.533238 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-94gfd" event={"ID":"3696cad8-c9be-4efd-982a-a4c1d6de858b","Type":"ContainerDied","Data":"b919558b67bf3e2ddc28345c59d249643aab8dfa5bcabcfff0dedaa84cb15193"} Nov 23 06:58:56 crc kubenswrapper[4559]: I1123 06:58:56.533252 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-94gfd" Nov 23 06:58:56 crc kubenswrapper[4559]: I1123 06:58:56.533260 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b919558b67bf3e2ddc28345c59d249643aab8dfa5bcabcfff0dedaa84cb15193" Nov 23 06:58:56 crc kubenswrapper[4559]: I1123 06:58:56.535461 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b8455895f-w72ch" event={"ID":"4460f104-a577-49e6-bb5a-dc50cb3401eb","Type":"ContainerDied","Data":"e8d3d48cc45ee9c94b55291517a0d9e1e1e37f3f67309e35f84fd45f8abf570f"} Nov 23 06:58:56 crc kubenswrapper[4559]: I1123 06:58:56.535553 4559 scope.go:117] "RemoveContainer" containerID="a0141bb4b5a5004234d5f211f1fb42f3f155c25253aa7cce564a0e938e8bc619" Nov 23 06:58:56 crc kubenswrapper[4559]: I1123 06:58:56.535778 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b8455895f-w72ch" Nov 23 06:58:56 crc kubenswrapper[4559]: I1123 06:58:56.536045 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7783a8fe-c77b-4e05-8c52-3e11d3a92196-etc-swift\") pod \"swift-storage-0\" (UID: \"7783a8fe-c77b-4e05-8c52-3e11d3a92196\") " pod="openstack/swift-storage-0" Nov 23 06:58:56 crc kubenswrapper[4559]: I1123 06:58:56.538222 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-blj85-config-mrnpd" event={"ID":"0e7f2ae1-1734-49cb-998e-efe380cc6c47","Type":"ContainerDied","Data":"b164ca75df46140716aa0764387fb5fb843d2da2c97a5864be5e853f74cfd5f5"} Nov 23 06:58:56 crc kubenswrapper[4559]: I1123 06:58:56.538258 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b164ca75df46140716aa0764387fb5fb843d2da2c97a5864be5e853f74cfd5f5" Nov 23 06:58:56 crc kubenswrapper[4559]: I1123 06:58:56.538271 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-blj85-config-mrnpd" Nov 23 06:58:56 crc kubenswrapper[4559]: I1123 06:58:56.549054 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-dwbhb" podStartSLOduration=2.239472255 podStartE2EDuration="18.549042119s" podCreationTimestamp="2025-11-23 06:58:38 +0000 UTC" firstStartedPulling="2025-11-23 06:58:39.258411827 +0000 UTC m=+821.280397440" lastFinishedPulling="2025-11-23 06:58:55.567981689 +0000 UTC m=+837.589967304" observedRunningTime="2025-11-23 06:58:56.54616015 +0000 UTC m=+838.568145764" watchObservedRunningTime="2025-11-23 06:58:56.549042119 +0000 UTC m=+838.571027733" Nov 23 06:58:56 crc kubenswrapper[4559]: I1123 06:58:56.593117 4559 scope.go:117] "RemoveContainer" containerID="d656a10bef82c04d5729a7b40002f50b4178794e85a139ecb2cbf8435b833736" Nov 23 06:58:56 crc kubenswrapper[4559]: I1123 06:58:56.600959 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.525145136 podStartE2EDuration="1m14.600942122s" podCreationTimestamp="2025-11-23 06:57:42 +0000 UTC" firstStartedPulling="2025-11-23 06:57:44.70894626 +0000 UTC m=+766.730931873" lastFinishedPulling="2025-11-23 06:58:20.784743245 +0000 UTC m=+802.806728859" observedRunningTime="2025-11-23 06:58:56.571015718 +0000 UTC m=+838.593001333" watchObservedRunningTime="2025-11-23 06:58:56.600942122 +0000 UTC m=+838.622927736" Nov 23 06:58:56 crc kubenswrapper[4559]: I1123 06:58:56.606703 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b8455895f-w72ch"] Nov 23 06:58:56 crc kubenswrapper[4559]: I1123 06:58:56.612111 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b8455895f-w72ch"] Nov 23 06:58:56 crc kubenswrapper[4559]: I1123 06:58:56.757462 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 23 06:58:56 crc kubenswrapper[4559]: I1123 06:58:56.788391 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-blj85-config-mrnpd"] Nov 23 06:58:56 crc kubenswrapper[4559]: I1123 06:58:56.790601 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-blj85-config-mrnpd"] Nov 23 06:58:57 crc kubenswrapper[4559]: I1123 06:58:57.235221 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 23 06:58:57 crc kubenswrapper[4559]: W1123 06:58:57.256595 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7783a8fe_c77b_4e05_8c52_3e11d3a92196.slice/crio-2c67ebcf3579ca37072f68340f807c786eb2f6905f4daff08fb629e066abe9db WatchSource:0}: Error finding container 2c67ebcf3579ca37072f68340f807c786eb2f6905f4daff08fb629e066abe9db: Status 404 returned error can't find the container with id 2c67ebcf3579ca37072f68340f807c786eb2f6905f4daff08fb629e066abe9db Nov 23 06:58:57 crc kubenswrapper[4559]: I1123 06:58:57.546845 4559 generic.go:334] "Generic (PLEG): container finished" podID="61e2f87a-db68-432a-8d13-94087d1afb17" containerID="3ed96cde7ed68fbff6a9ee5c583b3e101eecb01f52f5e3d5f2ca7b0c38bf7687" exitCode=0 Nov 23 06:58:57 crc kubenswrapper[4559]: I1123 06:58:57.546933 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"61e2f87a-db68-432a-8d13-94087d1afb17","Type":"ContainerDied","Data":"3ed96cde7ed68fbff6a9ee5c583b3e101eecb01f52f5e3d5f2ca7b0c38bf7687"} Nov 23 06:58:57 crc kubenswrapper[4559]: I1123 06:58:57.550499 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7783a8fe-c77b-4e05-8c52-3e11d3a92196","Type":"ContainerStarted","Data":"2c67ebcf3579ca37072f68340f807c786eb2f6905f4daff08fb629e066abe9db"} Nov 23 06:58:58 crc kubenswrapper[4559]: I1123 06:58:58.283260 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e7f2ae1-1734-49cb-998e-efe380cc6c47" path="/var/lib/kubelet/pods/0e7f2ae1-1734-49cb-998e-efe380cc6c47/volumes" Nov 23 06:58:58 crc kubenswrapper[4559]: I1123 06:58:58.284789 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4460f104-a577-49e6-bb5a-dc50cb3401eb" path="/var/lib/kubelet/pods/4460f104-a577-49e6-bb5a-dc50cb3401eb/volumes" Nov 23 06:58:58 crc kubenswrapper[4559]: I1123 06:58:58.563182 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"61e2f87a-db68-432a-8d13-94087d1afb17","Type":"ContainerStarted","Data":"94b662f9f4f42e326fff684a3b4eeacd903e6c78f20d5f630e934ee6cb875106"} Nov 23 06:58:58 crc kubenswrapper[4559]: I1123 06:58:58.564053 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:58:58 crc kubenswrapper[4559]: I1123 06:58:58.585181 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=-9223371961.269608 podStartE2EDuration="1m15.585168558s" podCreationTimestamp="2025-11-23 06:57:43 +0000 UTC" firstStartedPulling="2025-11-23 06:57:45.02522964 +0000 UTC m=+767.047215253" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:58:58.581743838 +0000 UTC m=+840.603729452" watchObservedRunningTime="2025-11-23 06:58:58.585168558 +0000 UTC m=+840.607154162" Nov 23 06:58:59 crc kubenswrapper[4559]: I1123 06:58:59.572369 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7783a8fe-c77b-4e05-8c52-3e11d3a92196","Type":"ContainerStarted","Data":"70a62a15ba669b4b3e8753b608e24f921f356f055fcff839b6c2a563d0a26845"} Nov 23 06:58:59 crc kubenswrapper[4559]: I1123 06:58:59.572572 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7783a8fe-c77b-4e05-8c52-3e11d3a92196","Type":"ContainerStarted","Data":"6e756539afa922e7198d68319dd70c3c53367ad00eaa30b5befc582db1a5fc33"} Nov 23 06:58:59 crc kubenswrapper[4559]: I1123 06:58:59.572584 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7783a8fe-c77b-4e05-8c52-3e11d3a92196","Type":"ContainerStarted","Data":"825a7e85288a2ccb28ae0f3b3c3a36b48ecb07809c65d4b520e2916d12f90e74"} Nov 23 06:58:59 crc kubenswrapper[4559]: I1123 06:58:59.572592 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7783a8fe-c77b-4e05-8c52-3e11d3a92196","Type":"ContainerStarted","Data":"4b3406ffc8bd65b85e37d0c81872269af766d5e5aef2830f3c25358b5f1ee017"} Nov 23 06:59:00 crc kubenswrapper[4559]: I1123 06:59:00.579512 4559 generic.go:334] "Generic (PLEG): container finished" podID="a5cc366d-e466-476d-a318-1582a4de4a93" containerID="28c045ec04f1f1fbec6423600a427ac6ae17f3d9f67ebec610f72f1a248b0593" exitCode=0 Nov 23 06:59:00 crc kubenswrapper[4559]: I1123 06:59:00.579553 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-dwbhb" event={"ID":"a5cc366d-e466-476d-a318-1582a4de4a93","Type":"ContainerDied","Data":"28c045ec04f1f1fbec6423600a427ac6ae17f3d9f67ebec610f72f1a248b0593"} Nov 23 06:59:01 crc kubenswrapper[4559]: I1123 06:59:01.925079 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-dwbhb" Nov 23 06:59:02 crc kubenswrapper[4559]: I1123 06:59:02.117111 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a5cc366d-e466-476d-a318-1582a4de4a93-db-sync-config-data\") pod \"a5cc366d-e466-476d-a318-1582a4de4a93\" (UID: \"a5cc366d-e466-476d-a318-1582a4de4a93\") " Nov 23 06:59:02 crc kubenswrapper[4559]: I1123 06:59:02.117175 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5cc366d-e466-476d-a318-1582a4de4a93-config-data\") pod \"a5cc366d-e466-476d-a318-1582a4de4a93\" (UID: \"a5cc366d-e466-476d-a318-1582a4de4a93\") " Nov 23 06:59:02 crc kubenswrapper[4559]: I1123 06:59:02.117202 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cc366d-e466-476d-a318-1582a4de4a93-combined-ca-bundle\") pod \"a5cc366d-e466-476d-a318-1582a4de4a93\" (UID: \"a5cc366d-e466-476d-a318-1582a4de4a93\") " Nov 23 06:59:02 crc kubenswrapper[4559]: I1123 06:59:02.117356 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58t7r\" (UniqueName: \"kubernetes.io/projected/a5cc366d-e466-476d-a318-1582a4de4a93-kube-api-access-58t7r\") pod \"a5cc366d-e466-476d-a318-1582a4de4a93\" (UID: \"a5cc366d-e466-476d-a318-1582a4de4a93\") " Nov 23 06:59:02 crc kubenswrapper[4559]: I1123 06:59:02.122688 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5cc366d-e466-476d-a318-1582a4de4a93-kube-api-access-58t7r" (OuterVolumeSpecName: "kube-api-access-58t7r") pod "a5cc366d-e466-476d-a318-1582a4de4a93" (UID: "a5cc366d-e466-476d-a318-1582a4de4a93"). InnerVolumeSpecName "kube-api-access-58t7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:59:02 crc kubenswrapper[4559]: I1123 06:59:02.123303 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5cc366d-e466-476d-a318-1582a4de4a93-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a5cc366d-e466-476d-a318-1582a4de4a93" (UID: "a5cc366d-e466-476d-a318-1582a4de4a93"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:02 crc kubenswrapper[4559]: I1123 06:59:02.136245 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5cc366d-e466-476d-a318-1582a4de4a93-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5cc366d-e466-476d-a318-1582a4de4a93" (UID: "a5cc366d-e466-476d-a318-1582a4de4a93"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:02 crc kubenswrapper[4559]: I1123 06:59:02.147576 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5cc366d-e466-476d-a318-1582a4de4a93-config-data" (OuterVolumeSpecName: "config-data") pod "a5cc366d-e466-476d-a318-1582a4de4a93" (UID: "a5cc366d-e466-476d-a318-1582a4de4a93"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:02 crc kubenswrapper[4559]: I1123 06:59:02.218927 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58t7r\" (UniqueName: \"kubernetes.io/projected/a5cc366d-e466-476d-a318-1582a4de4a93-kube-api-access-58t7r\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:02 crc kubenswrapper[4559]: I1123 06:59:02.218953 4559 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a5cc366d-e466-476d-a318-1582a4de4a93-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:02 crc kubenswrapper[4559]: I1123 06:59:02.218961 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5cc366d-e466-476d-a318-1582a4de4a93-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:02 crc kubenswrapper[4559]: I1123 06:59:02.218970 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cc366d-e466-476d-a318-1582a4de4a93-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:02 crc kubenswrapper[4559]: I1123 06:59:02.592549 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-dwbhb" event={"ID":"a5cc366d-e466-476d-a318-1582a4de4a93","Type":"ContainerDied","Data":"79c52b10f05d82764a475ebd15cc13f73d77ad2e6226ec8c24c50873d1710b4f"} Nov 23 06:59:02 crc kubenswrapper[4559]: I1123 06:59:02.592592 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79c52b10f05d82764a475ebd15cc13f73d77ad2e6226ec8c24c50873d1710b4f" Nov 23 06:59:02 crc kubenswrapper[4559]: I1123 06:59:02.592594 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-dwbhb" Nov 23 06:59:02 crc kubenswrapper[4559]: I1123 06:59:02.936570 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89d747df-9zxnl"] Nov 23 06:59:02 crc kubenswrapper[4559]: E1123 06:59:02.937550 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4460f104-a577-49e6-bb5a-dc50cb3401eb" containerName="dnsmasq-dns" Nov 23 06:59:02 crc kubenswrapper[4559]: I1123 06:59:02.937665 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="4460f104-a577-49e6-bb5a-dc50cb3401eb" containerName="dnsmasq-dns" Nov 23 06:59:02 crc kubenswrapper[4559]: E1123 06:59:02.937739 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5cc366d-e466-476d-a318-1582a4de4a93" containerName="glance-db-sync" Nov 23 06:59:02 crc kubenswrapper[4559]: I1123 06:59:02.937807 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5cc366d-e466-476d-a318-1582a4de4a93" containerName="glance-db-sync" Nov 23 06:59:02 crc kubenswrapper[4559]: E1123 06:59:02.937865 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3696cad8-c9be-4efd-982a-a4c1d6de858b" containerName="swift-ring-rebalance" Nov 23 06:59:02 crc kubenswrapper[4559]: I1123 06:59:02.937925 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="3696cad8-c9be-4efd-982a-a4c1d6de858b" containerName="swift-ring-rebalance" Nov 23 06:59:02 crc kubenswrapper[4559]: E1123 06:59:02.937997 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4460f104-a577-49e6-bb5a-dc50cb3401eb" containerName="init" Nov 23 06:59:02 crc kubenswrapper[4559]: I1123 06:59:02.938045 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="4460f104-a577-49e6-bb5a-dc50cb3401eb" containerName="init" Nov 23 06:59:02 crc kubenswrapper[4559]: E1123 06:59:02.938103 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e7f2ae1-1734-49cb-998e-efe380cc6c47" containerName="ovn-config" Nov 23 06:59:02 crc kubenswrapper[4559]: I1123 06:59:02.938147 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e7f2ae1-1734-49cb-998e-efe380cc6c47" containerName="ovn-config" Nov 23 06:59:02 crc kubenswrapper[4559]: I1123 06:59:02.938372 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e7f2ae1-1734-49cb-998e-efe380cc6c47" containerName="ovn-config" Nov 23 06:59:02 crc kubenswrapper[4559]: I1123 06:59:02.938426 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="3696cad8-c9be-4efd-982a-a4c1d6de858b" containerName="swift-ring-rebalance" Nov 23 06:59:02 crc kubenswrapper[4559]: I1123 06:59:02.938502 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5cc366d-e466-476d-a318-1582a4de4a93" containerName="glance-db-sync" Nov 23 06:59:02 crc kubenswrapper[4559]: I1123 06:59:02.938558 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="4460f104-a577-49e6-bb5a-dc50cb3401eb" containerName="dnsmasq-dns" Nov 23 06:59:02 crc kubenswrapper[4559]: I1123 06:59:02.939412 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89d747df-9zxnl" Nov 23 06:59:02 crc kubenswrapper[4559]: I1123 06:59:02.952026 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89d747df-9zxnl"] Nov 23 06:59:03 crc kubenswrapper[4559]: I1123 06:59:03.134342 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-dns-svc\") pod \"dnsmasq-dns-89d747df-9zxnl\" (UID: \"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5\") " pod="openstack/dnsmasq-dns-89d747df-9zxnl" Nov 23 06:59:03 crc kubenswrapper[4559]: I1123 06:59:03.134717 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-ovsdbserver-sb\") pod \"dnsmasq-dns-89d747df-9zxnl\" (UID: \"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5\") " pod="openstack/dnsmasq-dns-89d747df-9zxnl" Nov 23 06:59:03 crc kubenswrapper[4559]: I1123 06:59:03.134865 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-config\") pod \"dnsmasq-dns-89d747df-9zxnl\" (UID: \"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5\") " pod="openstack/dnsmasq-dns-89d747df-9zxnl" Nov 23 06:59:03 crc kubenswrapper[4559]: I1123 06:59:03.135090 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-ovsdbserver-nb\") pod \"dnsmasq-dns-89d747df-9zxnl\" (UID: \"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5\") " pod="openstack/dnsmasq-dns-89d747df-9zxnl" Nov 23 06:59:03 crc kubenswrapper[4559]: I1123 06:59:03.135264 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49mrl\" (UniqueName: \"kubernetes.io/projected/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-kube-api-access-49mrl\") pod \"dnsmasq-dns-89d747df-9zxnl\" (UID: \"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5\") " pod="openstack/dnsmasq-dns-89d747df-9zxnl" Nov 23 06:59:03 crc kubenswrapper[4559]: I1123 06:59:03.236565 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-ovsdbserver-nb\") pod \"dnsmasq-dns-89d747df-9zxnl\" (UID: \"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5\") " pod="openstack/dnsmasq-dns-89d747df-9zxnl" Nov 23 06:59:03 crc kubenswrapper[4559]: I1123 06:59:03.236725 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49mrl\" (UniqueName: \"kubernetes.io/projected/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-kube-api-access-49mrl\") pod \"dnsmasq-dns-89d747df-9zxnl\" (UID: \"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5\") " pod="openstack/dnsmasq-dns-89d747df-9zxnl" Nov 23 06:59:03 crc kubenswrapper[4559]: I1123 06:59:03.236746 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-dns-svc\") pod \"dnsmasq-dns-89d747df-9zxnl\" (UID: \"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5\") " pod="openstack/dnsmasq-dns-89d747df-9zxnl" Nov 23 06:59:03 crc kubenswrapper[4559]: I1123 06:59:03.236776 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-ovsdbserver-sb\") pod \"dnsmasq-dns-89d747df-9zxnl\" (UID: \"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5\") " pod="openstack/dnsmasq-dns-89d747df-9zxnl" Nov 23 06:59:03 crc kubenswrapper[4559]: I1123 06:59:03.236818 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-config\") pod \"dnsmasq-dns-89d747df-9zxnl\" (UID: \"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5\") " pod="openstack/dnsmasq-dns-89d747df-9zxnl" Nov 23 06:59:03 crc kubenswrapper[4559]: I1123 06:59:03.237502 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-ovsdbserver-nb\") pod \"dnsmasq-dns-89d747df-9zxnl\" (UID: \"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5\") " pod="openstack/dnsmasq-dns-89d747df-9zxnl" Nov 23 06:59:03 crc kubenswrapper[4559]: I1123 06:59:03.237690 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-config\") pod \"dnsmasq-dns-89d747df-9zxnl\" (UID: \"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5\") " pod="openstack/dnsmasq-dns-89d747df-9zxnl" Nov 23 06:59:03 crc kubenswrapper[4559]: I1123 06:59:03.238619 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-dns-svc\") pod \"dnsmasq-dns-89d747df-9zxnl\" (UID: \"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5\") " pod="openstack/dnsmasq-dns-89d747df-9zxnl" Nov 23 06:59:03 crc kubenswrapper[4559]: I1123 06:59:03.238924 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-ovsdbserver-sb\") pod \"dnsmasq-dns-89d747df-9zxnl\" (UID: \"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5\") " pod="openstack/dnsmasq-dns-89d747df-9zxnl" Nov 23 06:59:03 crc kubenswrapper[4559]: I1123 06:59:03.257192 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49mrl\" (UniqueName: \"kubernetes.io/projected/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-kube-api-access-49mrl\") pod \"dnsmasq-dns-89d747df-9zxnl\" (UID: \"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5\") " pod="openstack/dnsmasq-dns-89d747df-9zxnl" Nov 23 06:59:03 crc kubenswrapper[4559]: I1123 06:59:03.553192 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89d747df-9zxnl" Nov 23 06:59:03 crc kubenswrapper[4559]: I1123 06:59:03.621077 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7783a8fe-c77b-4e05-8c52-3e11d3a92196","Type":"ContainerStarted","Data":"520924d7682b7ae9ac6aeeb61f132fed31fffac5f603da769b8c0a510538ae54"} Nov 23 06:59:03 crc kubenswrapper[4559]: I1123 06:59:03.621133 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7783a8fe-c77b-4e05-8c52-3e11d3a92196","Type":"ContainerStarted","Data":"78da37ff07422cf9fa4ba258de57d4d099a056677c328030c688b0a23add984e"} Nov 23 06:59:04 crc kubenswrapper[4559]: I1123 06:59:04.035792 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89d747df-9zxnl"] Nov 23 06:59:04 crc kubenswrapper[4559]: W1123 06:59:04.037447 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ca808ef_7d2e_4df7_9418_6c94f8bb22e5.slice/crio-8b9003ba5c8334bd6826d118aa5e8eeff64586dd06b2ea5ab47db3411d376435 WatchSource:0}: Error finding container 8b9003ba5c8334bd6826d118aa5e8eeff64586dd06b2ea5ab47db3411d376435: Status 404 returned error can't find the container with id 8b9003ba5c8334bd6826d118aa5e8eeff64586dd06b2ea5ab47db3411d376435 Nov 23 06:59:04 crc kubenswrapper[4559]: I1123 06:59:04.630124 4559 generic.go:334] "Generic (PLEG): container finished" podID="0ca808ef-7d2e-4df7-9418-6c94f8bb22e5" containerID="35d7af2a2ca38cc806e5d097eb5282624f56ed4479dba807942b71399216d6e4" exitCode=0 Nov 23 06:59:04 crc kubenswrapper[4559]: I1123 06:59:04.630167 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89d747df-9zxnl" event={"ID":"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5","Type":"ContainerDied","Data":"35d7af2a2ca38cc806e5d097eb5282624f56ed4479dba807942b71399216d6e4"} Nov 23 06:59:04 crc kubenswrapper[4559]: I1123 06:59:04.630418 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89d747df-9zxnl" event={"ID":"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5","Type":"ContainerStarted","Data":"8b9003ba5c8334bd6826d118aa5e8eeff64586dd06b2ea5ab47db3411d376435"} Nov 23 06:59:04 crc kubenswrapper[4559]: I1123 06:59:04.636198 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7783a8fe-c77b-4e05-8c52-3e11d3a92196","Type":"ContainerStarted","Data":"524d2874434edf9c532b6d44e6fdadde3d606ff39cd9caa6dcf60e5c461f5454"} Nov 23 06:59:04 crc kubenswrapper[4559]: I1123 06:59:04.636240 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7783a8fe-c77b-4e05-8c52-3e11d3a92196","Type":"ContainerStarted","Data":"2ce3c80ccff8b27fe6a888177bc16859bb4c3fde52a7b300c1eb1387e669c7be"} Nov 23 06:59:05 crc kubenswrapper[4559]: I1123 06:59:05.648982 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7783a8fe-c77b-4e05-8c52-3e11d3a92196","Type":"ContainerStarted","Data":"df0ec37b61958b04315f95b3c49d1efce6057fee18ffe27f6d5ec7dc282249d2"} Nov 23 06:59:05 crc kubenswrapper[4559]: I1123 06:59:05.649203 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7783a8fe-c77b-4e05-8c52-3e11d3a92196","Type":"ContainerStarted","Data":"d644b9d0647022f91096e8ce9889bfaacf8d32eaaccb33bd00d1e7451d5518f1"} Nov 23 06:59:05 crc kubenswrapper[4559]: I1123 06:59:05.649213 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7783a8fe-c77b-4e05-8c52-3e11d3a92196","Type":"ContainerStarted","Data":"d70225456c1f43fe567c7666a2c6a45d89168995e179bf2cd16028ec6875e91e"} Nov 23 06:59:05 crc kubenswrapper[4559]: I1123 06:59:05.652544 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89d747df-9zxnl" event={"ID":"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5","Type":"ContainerStarted","Data":"8e401643573ecab519064ebaa249f51ea1295b7156e39d4b5c5448b36b5769c5"} Nov 23 06:59:05 crc kubenswrapper[4559]: I1123 06:59:05.652801 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89d747df-9zxnl" Nov 23 06:59:05 crc kubenswrapper[4559]: I1123 06:59:05.676058 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89d747df-9zxnl" podStartSLOduration=3.676043275 podStartE2EDuration="3.676043275s" podCreationTimestamp="2025-11-23 06:59:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:59:05.669104866 +0000 UTC m=+847.691090480" watchObservedRunningTime="2025-11-23 06:59:05.676043275 +0000 UTC m=+847.698028899" Nov 23 06:59:06 crc kubenswrapper[4559]: I1123 06:59:06.664774 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7783a8fe-c77b-4e05-8c52-3e11d3a92196","Type":"ContainerStarted","Data":"e0ef79e96412af97d4bc84d55de7f9a6566fa2ffb201413e30130018fd3603fc"} Nov 23 06:59:06 crc kubenswrapper[4559]: I1123 06:59:06.665000 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7783a8fe-c77b-4e05-8c52-3e11d3a92196","Type":"ContainerStarted","Data":"0a9c8c8bc52ec1a76404f25d3e033b071564d22494c6483b3d274b1de22341a7"} Nov 23 06:59:06 crc kubenswrapper[4559]: I1123 06:59:06.665012 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7783a8fe-c77b-4e05-8c52-3e11d3a92196","Type":"ContainerStarted","Data":"ca8b59bb33eff2705fa4c6e9f14e09847ce2ed731ede3282e1ecb650d366a0f9"} Nov 23 06:59:06 crc kubenswrapper[4559]: I1123 06:59:06.665022 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7783a8fe-c77b-4e05-8c52-3e11d3a92196","Type":"ContainerStarted","Data":"79a4525e449a45daf4aa33edf839f2238f0ec6b50c4ebace65fc377ed62faf0f"} Nov 23 06:59:06 crc kubenswrapper[4559]: I1123 06:59:06.693808 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=19.762082453 podStartE2EDuration="27.693786226s" podCreationTimestamp="2025-11-23 06:58:39 +0000 UTC" firstStartedPulling="2025-11-23 06:58:57.267443594 +0000 UTC m=+839.289429207" lastFinishedPulling="2025-11-23 06:59:05.199147366 +0000 UTC m=+847.221132980" observedRunningTime="2025-11-23 06:59:06.688045029 +0000 UTC m=+848.710030643" watchObservedRunningTime="2025-11-23 06:59:06.693786226 +0000 UTC m=+848.715771840" Nov 23 06:59:06 crc kubenswrapper[4559]: I1123 06:59:06.910596 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89d747df-9zxnl"] Nov 23 06:59:06 crc kubenswrapper[4559]: I1123 06:59:06.941694 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-648ff47655-pcz6l"] Nov 23 06:59:06 crc kubenswrapper[4559]: I1123 06:59:06.943060 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-648ff47655-pcz6l" Nov 23 06:59:06 crc kubenswrapper[4559]: I1123 06:59:06.947398 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 23 06:59:06 crc kubenswrapper[4559]: I1123 06:59:06.973968 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-648ff47655-pcz6l"] Nov 23 06:59:07 crc kubenswrapper[4559]: I1123 06:59:07.099201 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-dns-svc\") pod \"dnsmasq-dns-648ff47655-pcz6l\" (UID: \"9dafbd81-21bc-4d7e-a61d-ba694502490c\") " pod="openstack/dnsmasq-dns-648ff47655-pcz6l" Nov 23 06:59:07 crc kubenswrapper[4559]: I1123 06:59:07.099482 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-dns-swift-storage-0\") pod \"dnsmasq-dns-648ff47655-pcz6l\" (UID: \"9dafbd81-21bc-4d7e-a61d-ba694502490c\") " pod="openstack/dnsmasq-dns-648ff47655-pcz6l" Nov 23 06:59:07 crc kubenswrapper[4559]: I1123 06:59:07.099708 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rj6rh\" (UniqueName: \"kubernetes.io/projected/9dafbd81-21bc-4d7e-a61d-ba694502490c-kube-api-access-rj6rh\") pod \"dnsmasq-dns-648ff47655-pcz6l\" (UID: \"9dafbd81-21bc-4d7e-a61d-ba694502490c\") " pod="openstack/dnsmasq-dns-648ff47655-pcz6l" Nov 23 06:59:07 crc kubenswrapper[4559]: I1123 06:59:07.099817 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-ovsdbserver-nb\") pod \"dnsmasq-dns-648ff47655-pcz6l\" (UID: \"9dafbd81-21bc-4d7e-a61d-ba694502490c\") " pod="openstack/dnsmasq-dns-648ff47655-pcz6l" Nov 23 06:59:07 crc kubenswrapper[4559]: I1123 06:59:07.099981 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-config\") pod \"dnsmasq-dns-648ff47655-pcz6l\" (UID: \"9dafbd81-21bc-4d7e-a61d-ba694502490c\") " pod="openstack/dnsmasq-dns-648ff47655-pcz6l" Nov 23 06:59:07 crc kubenswrapper[4559]: I1123 06:59:07.100159 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-ovsdbserver-sb\") pod \"dnsmasq-dns-648ff47655-pcz6l\" (UID: \"9dafbd81-21bc-4d7e-a61d-ba694502490c\") " pod="openstack/dnsmasq-dns-648ff47655-pcz6l" Nov 23 06:59:07 crc kubenswrapper[4559]: I1123 06:59:07.201618 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-dns-svc\") pod \"dnsmasq-dns-648ff47655-pcz6l\" (UID: \"9dafbd81-21bc-4d7e-a61d-ba694502490c\") " pod="openstack/dnsmasq-dns-648ff47655-pcz6l" Nov 23 06:59:07 crc kubenswrapper[4559]: I1123 06:59:07.201889 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-dns-swift-storage-0\") pod \"dnsmasq-dns-648ff47655-pcz6l\" (UID: \"9dafbd81-21bc-4d7e-a61d-ba694502490c\") " pod="openstack/dnsmasq-dns-648ff47655-pcz6l" Nov 23 06:59:07 crc kubenswrapper[4559]: I1123 06:59:07.202023 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rj6rh\" (UniqueName: \"kubernetes.io/projected/9dafbd81-21bc-4d7e-a61d-ba694502490c-kube-api-access-rj6rh\") pod \"dnsmasq-dns-648ff47655-pcz6l\" (UID: \"9dafbd81-21bc-4d7e-a61d-ba694502490c\") " pod="openstack/dnsmasq-dns-648ff47655-pcz6l" Nov 23 06:59:07 crc kubenswrapper[4559]: I1123 06:59:07.202103 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-ovsdbserver-nb\") pod \"dnsmasq-dns-648ff47655-pcz6l\" (UID: \"9dafbd81-21bc-4d7e-a61d-ba694502490c\") " pod="openstack/dnsmasq-dns-648ff47655-pcz6l" Nov 23 06:59:07 crc kubenswrapper[4559]: I1123 06:59:07.202187 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-config\") pod \"dnsmasq-dns-648ff47655-pcz6l\" (UID: \"9dafbd81-21bc-4d7e-a61d-ba694502490c\") " pod="openstack/dnsmasq-dns-648ff47655-pcz6l" Nov 23 06:59:07 crc kubenswrapper[4559]: I1123 06:59:07.202278 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-ovsdbserver-sb\") pod \"dnsmasq-dns-648ff47655-pcz6l\" (UID: \"9dafbd81-21bc-4d7e-a61d-ba694502490c\") " pod="openstack/dnsmasq-dns-648ff47655-pcz6l" Nov 23 06:59:07 crc kubenswrapper[4559]: I1123 06:59:07.202743 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-dns-swift-storage-0\") pod \"dnsmasq-dns-648ff47655-pcz6l\" (UID: \"9dafbd81-21bc-4d7e-a61d-ba694502490c\") " pod="openstack/dnsmasq-dns-648ff47655-pcz6l" Nov 23 06:59:07 crc kubenswrapper[4559]: I1123 06:59:07.203047 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-ovsdbserver-nb\") pod \"dnsmasq-dns-648ff47655-pcz6l\" (UID: \"9dafbd81-21bc-4d7e-a61d-ba694502490c\") " pod="openstack/dnsmasq-dns-648ff47655-pcz6l" Nov 23 06:59:07 crc kubenswrapper[4559]: I1123 06:59:07.203139 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-ovsdbserver-sb\") pod \"dnsmasq-dns-648ff47655-pcz6l\" (UID: \"9dafbd81-21bc-4d7e-a61d-ba694502490c\") " pod="openstack/dnsmasq-dns-648ff47655-pcz6l" Nov 23 06:59:07 crc kubenswrapper[4559]: I1123 06:59:07.203488 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-dns-svc\") pod \"dnsmasq-dns-648ff47655-pcz6l\" (UID: \"9dafbd81-21bc-4d7e-a61d-ba694502490c\") " pod="openstack/dnsmasq-dns-648ff47655-pcz6l" Nov 23 06:59:07 crc kubenswrapper[4559]: I1123 06:59:07.203602 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-config\") pod \"dnsmasq-dns-648ff47655-pcz6l\" (UID: \"9dafbd81-21bc-4d7e-a61d-ba694502490c\") " pod="openstack/dnsmasq-dns-648ff47655-pcz6l" Nov 23 06:59:07 crc kubenswrapper[4559]: I1123 06:59:07.217665 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rj6rh\" (UniqueName: \"kubernetes.io/projected/9dafbd81-21bc-4d7e-a61d-ba694502490c-kube-api-access-rj6rh\") pod \"dnsmasq-dns-648ff47655-pcz6l\" (UID: \"9dafbd81-21bc-4d7e-a61d-ba694502490c\") " pod="openstack/dnsmasq-dns-648ff47655-pcz6l" Nov 23 06:59:07 crc kubenswrapper[4559]: I1123 06:59:07.282763 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-648ff47655-pcz6l" Nov 23 06:59:07 crc kubenswrapper[4559]: I1123 06:59:07.652346 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-648ff47655-pcz6l"] Nov 23 06:59:07 crc kubenswrapper[4559]: W1123 06:59:07.655776 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9dafbd81_21bc_4d7e_a61d_ba694502490c.slice/crio-ed650f7549b264b84993bb17a274f71094c4cdd169ff3c9f95b62f50a41dd744 WatchSource:0}: Error finding container ed650f7549b264b84993bb17a274f71094c4cdd169ff3c9f95b62f50a41dd744: Status 404 returned error can't find the container with id ed650f7549b264b84993bb17a274f71094c4cdd169ff3c9f95b62f50a41dd744 Nov 23 06:59:07 crc kubenswrapper[4559]: I1123 06:59:07.673817 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-648ff47655-pcz6l" event={"ID":"9dafbd81-21bc-4d7e-a61d-ba694502490c","Type":"ContainerStarted","Data":"ed650f7549b264b84993bb17a274f71094c4cdd169ff3c9f95b62f50a41dd744"} Nov 23 06:59:07 crc kubenswrapper[4559]: I1123 06:59:07.673990 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89d747df-9zxnl" podUID="0ca808ef-7d2e-4df7-9418-6c94f8bb22e5" containerName="dnsmasq-dns" containerID="cri-o://8e401643573ecab519064ebaa249f51ea1295b7156e39d4b5c5448b36b5769c5" gracePeriod=10 Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.034999 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89d747df-9zxnl" Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.218911 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-config\") pod \"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5\" (UID: \"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5\") " Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.218959 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-ovsdbserver-sb\") pod \"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5\" (UID: \"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5\") " Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.219028 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-dns-svc\") pod \"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5\" (UID: \"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5\") " Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.219056 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49mrl\" (UniqueName: \"kubernetes.io/projected/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-kube-api-access-49mrl\") pod \"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5\" (UID: \"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5\") " Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.219086 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-ovsdbserver-nb\") pod \"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5\" (UID: \"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5\") " Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.222492 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-kube-api-access-49mrl" (OuterVolumeSpecName: "kube-api-access-49mrl") pod "0ca808ef-7d2e-4df7-9418-6c94f8bb22e5" (UID: "0ca808ef-7d2e-4df7-9418-6c94f8bb22e5"). InnerVolumeSpecName "kube-api-access-49mrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.247614 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0ca808ef-7d2e-4df7-9418-6c94f8bb22e5" (UID: "0ca808ef-7d2e-4df7-9418-6c94f8bb22e5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.248202 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0ca808ef-7d2e-4df7-9418-6c94f8bb22e5" (UID: "0ca808ef-7d2e-4df7-9418-6c94f8bb22e5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.250494 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-config" (OuterVolumeSpecName: "config") pod "0ca808ef-7d2e-4df7-9418-6c94f8bb22e5" (UID: "0ca808ef-7d2e-4df7-9418-6c94f8bb22e5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.252280 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0ca808ef-7d2e-4df7-9418-6c94f8bb22e5" (UID: "0ca808ef-7d2e-4df7-9418-6c94f8bb22e5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.321009 4559 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.321058 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49mrl\" (UniqueName: \"kubernetes.io/projected/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-kube-api-access-49mrl\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.321070 4559 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.321078 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.321087 4559 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.690221 4559 generic.go:334] "Generic (PLEG): container finished" podID="0ca808ef-7d2e-4df7-9418-6c94f8bb22e5" containerID="8e401643573ecab519064ebaa249f51ea1295b7156e39d4b5c5448b36b5769c5" exitCode=0 Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.690281 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89d747df-9zxnl" event={"ID":"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5","Type":"ContainerDied","Data":"8e401643573ecab519064ebaa249f51ea1295b7156e39d4b5c5448b36b5769c5"} Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.690313 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89d747df-9zxnl" event={"ID":"0ca808ef-7d2e-4df7-9418-6c94f8bb22e5","Type":"ContainerDied","Data":"8b9003ba5c8334bd6826d118aa5e8eeff64586dd06b2ea5ab47db3411d376435"} Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.690331 4559 scope.go:117] "RemoveContainer" containerID="8e401643573ecab519064ebaa249f51ea1295b7156e39d4b5c5448b36b5769c5" Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.690452 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89d747df-9zxnl" Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.692324 4559 generic.go:334] "Generic (PLEG): container finished" podID="9dafbd81-21bc-4d7e-a61d-ba694502490c" containerID="1978a78cbcd124df30e44e33995c35b8b2834890d4a10ee9c29996cf7b34b370" exitCode=0 Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.692344 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-648ff47655-pcz6l" event={"ID":"9dafbd81-21bc-4d7e-a61d-ba694502490c","Type":"ContainerDied","Data":"1978a78cbcd124df30e44e33995c35b8b2834890d4a10ee9c29996cf7b34b370"} Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.715863 4559 scope.go:117] "RemoveContainer" containerID="35d7af2a2ca38cc806e5d097eb5282624f56ed4479dba807942b71399216d6e4" Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.728969 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89d747df-9zxnl"] Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.734452 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89d747df-9zxnl"] Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.795969 4559 scope.go:117] "RemoveContainer" containerID="8e401643573ecab519064ebaa249f51ea1295b7156e39d4b5c5448b36b5769c5" Nov 23 06:59:08 crc kubenswrapper[4559]: E1123 06:59:08.796361 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e401643573ecab519064ebaa249f51ea1295b7156e39d4b5c5448b36b5769c5\": container with ID starting with 8e401643573ecab519064ebaa249f51ea1295b7156e39d4b5c5448b36b5769c5 not found: ID does not exist" containerID="8e401643573ecab519064ebaa249f51ea1295b7156e39d4b5c5448b36b5769c5" Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.796389 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e401643573ecab519064ebaa249f51ea1295b7156e39d4b5c5448b36b5769c5"} err="failed to get container status \"8e401643573ecab519064ebaa249f51ea1295b7156e39d4b5c5448b36b5769c5\": rpc error: code = NotFound desc = could not find container \"8e401643573ecab519064ebaa249f51ea1295b7156e39d4b5c5448b36b5769c5\": container with ID starting with 8e401643573ecab519064ebaa249f51ea1295b7156e39d4b5c5448b36b5769c5 not found: ID does not exist" Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.796409 4559 scope.go:117] "RemoveContainer" containerID="35d7af2a2ca38cc806e5d097eb5282624f56ed4479dba807942b71399216d6e4" Nov 23 06:59:08 crc kubenswrapper[4559]: E1123 06:59:08.796689 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35d7af2a2ca38cc806e5d097eb5282624f56ed4479dba807942b71399216d6e4\": container with ID starting with 35d7af2a2ca38cc806e5d097eb5282624f56ed4479dba807942b71399216d6e4 not found: ID does not exist" containerID="35d7af2a2ca38cc806e5d097eb5282624f56ed4479dba807942b71399216d6e4" Nov 23 06:59:08 crc kubenswrapper[4559]: I1123 06:59:08.796713 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35d7af2a2ca38cc806e5d097eb5282624f56ed4479dba807942b71399216d6e4"} err="failed to get container status \"35d7af2a2ca38cc806e5d097eb5282624f56ed4479dba807942b71399216d6e4\": rpc error: code = NotFound desc = could not find container \"35d7af2a2ca38cc806e5d097eb5282624f56ed4479dba807942b71399216d6e4\": container with ID starting with 35d7af2a2ca38cc806e5d097eb5282624f56ed4479dba807942b71399216d6e4 not found: ID does not exist" Nov 23 06:59:09 crc kubenswrapper[4559]: I1123 06:59:09.704268 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-648ff47655-pcz6l" event={"ID":"9dafbd81-21bc-4d7e-a61d-ba694502490c","Type":"ContainerStarted","Data":"b87622aa82b32ff182c836aa8a7f812dd11a95f19917e11e867c5184307dc0c6"} Nov 23 06:59:09 crc kubenswrapper[4559]: I1123 06:59:09.704657 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-648ff47655-pcz6l" Nov 23 06:59:09 crc kubenswrapper[4559]: I1123 06:59:09.718785 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-648ff47655-pcz6l" podStartSLOduration=3.71876014 podStartE2EDuration="3.71876014s" podCreationTimestamp="2025-11-23 06:59:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:59:09.716969593 +0000 UTC m=+851.738955207" watchObservedRunningTime="2025-11-23 06:59:09.71876014 +0000 UTC m=+851.740745755" Nov 23 06:59:10 crc kubenswrapper[4559]: I1123 06:59:10.281837 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ca808ef-7d2e-4df7-9418-6c94f8bb22e5" path="/var/lib/kubelet/pods/0ca808ef-7d2e-4df7-9418-6c94f8bb22e5/volumes" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.301829 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.519854 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-j4kkb"] Nov 23 06:59:14 crc kubenswrapper[4559]: E1123 06:59:14.520381 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ca808ef-7d2e-4df7-9418-6c94f8bb22e5" containerName="dnsmasq-dns" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.520403 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ca808ef-7d2e-4df7-9418-6c94f8bb22e5" containerName="dnsmasq-dns" Nov 23 06:59:14 crc kubenswrapper[4559]: E1123 06:59:14.520448 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ca808ef-7d2e-4df7-9418-6c94f8bb22e5" containerName="init" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.520454 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ca808ef-7d2e-4df7-9418-6c94f8bb22e5" containerName="init" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.520674 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ca808ef-7d2e-4df7-9418-6c94f8bb22e5" containerName="dnsmasq-dns" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.521918 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-j4kkb" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.540145 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-j4kkb"] Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.574700 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.623046 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-2xf6v"] Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.624079 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-2xf6v" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.637979 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-f4d5-account-create-q6dwz"] Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.638760 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f4d5-account-create-q6dwz" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.643161 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-2xf6v"] Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.643669 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.649664 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-f4d5-account-create-q6dwz"] Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.710552 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fstb\" (UniqueName: \"kubernetes.io/projected/39400620-4347-45f5-85ee-4bfed29cb5cd-kube-api-access-5fstb\") pod \"cinder-db-create-j4kkb\" (UID: \"39400620-4347-45f5-85ee-4bfed29cb5cd\") " pod="openstack/cinder-db-create-j4kkb" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.710794 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/39400620-4347-45f5-85ee-4bfed29cb5cd-operator-scripts\") pod \"cinder-db-create-j4kkb\" (UID: \"39400620-4347-45f5-85ee-4bfed29cb5cd\") " pod="openstack/cinder-db-create-j4kkb" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.736223 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-6ad9-account-create-bfsj7"] Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.737464 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6ad9-account-create-bfsj7" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.746209 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-6ad9-account-create-bfsj7"] Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.749747 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.812997 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/39400620-4347-45f5-85ee-4bfed29cb5cd-operator-scripts\") pod \"cinder-db-create-j4kkb\" (UID: \"39400620-4347-45f5-85ee-4bfed29cb5cd\") " pod="openstack/cinder-db-create-j4kkb" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.813311 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11264a11-4f18-4550-a335-c0fd08428786-operator-scripts\") pod \"cinder-f4d5-account-create-q6dwz\" (UID: \"11264a11-4f18-4550-a335-c0fd08428786\") " pod="openstack/cinder-f4d5-account-create-q6dwz" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.813449 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwl7d\" (UniqueName: \"kubernetes.io/projected/a2b9eb8d-f298-4085-9b0e-4c9c929e23c2-kube-api-access-rwl7d\") pod \"barbican-db-create-2xf6v\" (UID: \"a2b9eb8d-f298-4085-9b0e-4c9c929e23c2\") " pod="openstack/barbican-db-create-2xf6v" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.813680 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dt7mz\" (UniqueName: \"kubernetes.io/projected/11264a11-4f18-4550-a335-c0fd08428786-kube-api-access-dt7mz\") pod \"cinder-f4d5-account-create-q6dwz\" (UID: \"11264a11-4f18-4550-a335-c0fd08428786\") " pod="openstack/cinder-f4d5-account-create-q6dwz" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.813738 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/39400620-4347-45f5-85ee-4bfed29cb5cd-operator-scripts\") pod \"cinder-db-create-j4kkb\" (UID: \"39400620-4347-45f5-85ee-4bfed29cb5cd\") " pod="openstack/cinder-db-create-j4kkb" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.813814 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fstb\" (UniqueName: \"kubernetes.io/projected/39400620-4347-45f5-85ee-4bfed29cb5cd-kube-api-access-5fstb\") pod \"cinder-db-create-j4kkb\" (UID: \"39400620-4347-45f5-85ee-4bfed29cb5cd\") " pod="openstack/cinder-db-create-j4kkb" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.814186 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2b9eb8d-f298-4085-9b0e-4c9c929e23c2-operator-scripts\") pod \"barbican-db-create-2xf6v\" (UID: \"a2b9eb8d-f298-4085-9b0e-4c9c929e23c2\") " pod="openstack/barbican-db-create-2xf6v" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.823913 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-65wms"] Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.825182 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-65wms" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.833573 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fstb\" (UniqueName: \"kubernetes.io/projected/39400620-4347-45f5-85ee-4bfed29cb5cd-kube-api-access-5fstb\") pod \"cinder-db-create-j4kkb\" (UID: \"39400620-4347-45f5-85ee-4bfed29cb5cd\") " pod="openstack/cinder-db-create-j4kkb" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.835093 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-65wms"] Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.867202 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-j4kkb" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.882737 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-b6wwk"] Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.884143 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-b6wwk" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.887597 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.888926 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.889290 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.898329 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-b6wwk"] Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.901300 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xmhpd" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.915690 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2b9eb8d-f298-4085-9b0e-4c9c929e23c2-operator-scripts\") pod \"barbican-db-create-2xf6v\" (UID: \"a2b9eb8d-f298-4085-9b0e-4c9c929e23c2\") " pod="openstack/barbican-db-create-2xf6v" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.915751 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcd2k\" (UniqueName: \"kubernetes.io/projected/c6600577-55b0-439d-85c2-dab41fac1775-kube-api-access-fcd2k\") pod \"barbican-6ad9-account-create-bfsj7\" (UID: \"c6600577-55b0-439d-85c2-dab41fac1775\") " pod="openstack/barbican-6ad9-account-create-bfsj7" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.915788 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11264a11-4f18-4550-a335-c0fd08428786-operator-scripts\") pod \"cinder-f4d5-account-create-q6dwz\" (UID: \"11264a11-4f18-4550-a335-c0fd08428786\") " pod="openstack/cinder-f4d5-account-create-q6dwz" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.915813 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwl7d\" (UniqueName: \"kubernetes.io/projected/a2b9eb8d-f298-4085-9b0e-4c9c929e23c2-kube-api-access-rwl7d\") pod \"barbican-db-create-2xf6v\" (UID: \"a2b9eb8d-f298-4085-9b0e-4c9c929e23c2\") " pod="openstack/barbican-db-create-2xf6v" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.915834 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6600577-55b0-439d-85c2-dab41fac1775-operator-scripts\") pod \"barbican-6ad9-account-create-bfsj7\" (UID: \"c6600577-55b0-439d-85c2-dab41fac1775\") " pod="openstack/barbican-6ad9-account-create-bfsj7" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.915874 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dt7mz\" (UniqueName: \"kubernetes.io/projected/11264a11-4f18-4550-a335-c0fd08428786-kube-api-access-dt7mz\") pod \"cinder-f4d5-account-create-q6dwz\" (UID: \"11264a11-4f18-4550-a335-c0fd08428786\") " pod="openstack/cinder-f4d5-account-create-q6dwz" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.916675 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2b9eb8d-f298-4085-9b0e-4c9c929e23c2-operator-scripts\") pod \"barbican-db-create-2xf6v\" (UID: \"a2b9eb8d-f298-4085-9b0e-4c9c929e23c2\") " pod="openstack/barbican-db-create-2xf6v" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.917249 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11264a11-4f18-4550-a335-c0fd08428786-operator-scripts\") pod \"cinder-f4d5-account-create-q6dwz\" (UID: \"11264a11-4f18-4550-a335-c0fd08428786\") " pod="openstack/cinder-f4d5-account-create-q6dwz" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.938472 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dt7mz\" (UniqueName: \"kubernetes.io/projected/11264a11-4f18-4550-a335-c0fd08428786-kube-api-access-dt7mz\") pod \"cinder-f4d5-account-create-q6dwz\" (UID: \"11264a11-4f18-4550-a335-c0fd08428786\") " pod="openstack/cinder-f4d5-account-create-q6dwz" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.940220 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwl7d\" (UniqueName: \"kubernetes.io/projected/a2b9eb8d-f298-4085-9b0e-4c9c929e23c2-kube-api-access-rwl7d\") pod \"barbican-db-create-2xf6v\" (UID: \"a2b9eb8d-f298-4085-9b0e-4c9c929e23c2\") " pod="openstack/barbican-db-create-2xf6v" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.962082 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f4d5-account-create-q6dwz" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.971694 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-a32f-account-create-h2s4w"] Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.973137 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-a32f-account-create-h2s4w" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.974711 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 23 06:59:14 crc kubenswrapper[4559]: I1123 06:59:14.979157 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-a32f-account-create-h2s4w"] Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.017502 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6600577-55b0-439d-85c2-dab41fac1775-operator-scripts\") pod \"barbican-6ad9-account-create-bfsj7\" (UID: \"c6600577-55b0-439d-85c2-dab41fac1775\") " pod="openstack/barbican-6ad9-account-create-bfsj7" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.017667 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgwhn\" (UniqueName: \"kubernetes.io/projected/d08d731c-9663-4eb7-ac6d-8f500dc4b294-kube-api-access-fgwhn\") pod \"neutron-db-create-65wms\" (UID: \"d08d731c-9663-4eb7-ac6d-8f500dc4b294\") " pod="openstack/neutron-db-create-65wms" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.017698 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8cv2\" (UniqueName: \"kubernetes.io/projected/e43dc009-537b-4a2c-b481-e42f41eec38d-kube-api-access-f8cv2\") pod \"keystone-db-sync-b6wwk\" (UID: \"e43dc009-537b-4a2c-b481-e42f41eec38d\") " pod="openstack/keystone-db-sync-b6wwk" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.017744 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e43dc009-537b-4a2c-b481-e42f41eec38d-combined-ca-bundle\") pod \"keystone-db-sync-b6wwk\" (UID: \"e43dc009-537b-4a2c-b481-e42f41eec38d\") " pod="openstack/keystone-db-sync-b6wwk" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.017786 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d08d731c-9663-4eb7-ac6d-8f500dc4b294-operator-scripts\") pod \"neutron-db-create-65wms\" (UID: \"d08d731c-9663-4eb7-ac6d-8f500dc4b294\") " pod="openstack/neutron-db-create-65wms" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.017814 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e43dc009-537b-4a2c-b481-e42f41eec38d-config-data\") pod \"keystone-db-sync-b6wwk\" (UID: \"e43dc009-537b-4a2c-b481-e42f41eec38d\") " pod="openstack/keystone-db-sync-b6wwk" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.017853 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcd2k\" (UniqueName: \"kubernetes.io/projected/c6600577-55b0-439d-85c2-dab41fac1775-kube-api-access-fcd2k\") pod \"barbican-6ad9-account-create-bfsj7\" (UID: \"c6600577-55b0-439d-85c2-dab41fac1775\") " pod="openstack/barbican-6ad9-account-create-bfsj7" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.018185 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6600577-55b0-439d-85c2-dab41fac1775-operator-scripts\") pod \"barbican-6ad9-account-create-bfsj7\" (UID: \"c6600577-55b0-439d-85c2-dab41fac1775\") " pod="openstack/barbican-6ad9-account-create-bfsj7" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.032595 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcd2k\" (UniqueName: \"kubernetes.io/projected/c6600577-55b0-439d-85c2-dab41fac1775-kube-api-access-fcd2k\") pod \"barbican-6ad9-account-create-bfsj7\" (UID: \"c6600577-55b0-439d-85c2-dab41fac1775\") " pod="openstack/barbican-6ad9-account-create-bfsj7" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.052435 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6ad9-account-create-bfsj7" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.119780 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chxft\" (UniqueName: \"kubernetes.io/projected/229ac05d-35f6-4775-8298-c61aa1378406-kube-api-access-chxft\") pod \"neutron-a32f-account-create-h2s4w\" (UID: \"229ac05d-35f6-4775-8298-c61aa1378406\") " pod="openstack/neutron-a32f-account-create-h2s4w" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.120084 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgwhn\" (UniqueName: \"kubernetes.io/projected/d08d731c-9663-4eb7-ac6d-8f500dc4b294-kube-api-access-fgwhn\") pod \"neutron-db-create-65wms\" (UID: \"d08d731c-9663-4eb7-ac6d-8f500dc4b294\") " pod="openstack/neutron-db-create-65wms" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.120106 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8cv2\" (UniqueName: \"kubernetes.io/projected/e43dc009-537b-4a2c-b481-e42f41eec38d-kube-api-access-f8cv2\") pod \"keystone-db-sync-b6wwk\" (UID: \"e43dc009-537b-4a2c-b481-e42f41eec38d\") " pod="openstack/keystone-db-sync-b6wwk" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.120132 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/229ac05d-35f6-4775-8298-c61aa1378406-operator-scripts\") pod \"neutron-a32f-account-create-h2s4w\" (UID: \"229ac05d-35f6-4775-8298-c61aa1378406\") " pod="openstack/neutron-a32f-account-create-h2s4w" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.120182 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e43dc009-537b-4a2c-b481-e42f41eec38d-combined-ca-bundle\") pod \"keystone-db-sync-b6wwk\" (UID: \"e43dc009-537b-4a2c-b481-e42f41eec38d\") " pod="openstack/keystone-db-sync-b6wwk" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.120234 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d08d731c-9663-4eb7-ac6d-8f500dc4b294-operator-scripts\") pod \"neutron-db-create-65wms\" (UID: \"d08d731c-9663-4eb7-ac6d-8f500dc4b294\") " pod="openstack/neutron-db-create-65wms" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.120272 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e43dc009-537b-4a2c-b481-e42f41eec38d-config-data\") pod \"keystone-db-sync-b6wwk\" (UID: \"e43dc009-537b-4a2c-b481-e42f41eec38d\") " pod="openstack/keystone-db-sync-b6wwk" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.121112 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d08d731c-9663-4eb7-ac6d-8f500dc4b294-operator-scripts\") pod \"neutron-db-create-65wms\" (UID: \"d08d731c-9663-4eb7-ac6d-8f500dc4b294\") " pod="openstack/neutron-db-create-65wms" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.124245 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e43dc009-537b-4a2c-b481-e42f41eec38d-config-data\") pod \"keystone-db-sync-b6wwk\" (UID: \"e43dc009-537b-4a2c-b481-e42f41eec38d\") " pod="openstack/keystone-db-sync-b6wwk" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.124717 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e43dc009-537b-4a2c-b481-e42f41eec38d-combined-ca-bundle\") pod \"keystone-db-sync-b6wwk\" (UID: \"e43dc009-537b-4a2c-b481-e42f41eec38d\") " pod="openstack/keystone-db-sync-b6wwk" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.133986 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgwhn\" (UniqueName: \"kubernetes.io/projected/d08d731c-9663-4eb7-ac6d-8f500dc4b294-kube-api-access-fgwhn\") pod \"neutron-db-create-65wms\" (UID: \"d08d731c-9663-4eb7-ac6d-8f500dc4b294\") " pod="openstack/neutron-db-create-65wms" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.141860 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8cv2\" (UniqueName: \"kubernetes.io/projected/e43dc009-537b-4a2c-b481-e42f41eec38d-kube-api-access-f8cv2\") pod \"keystone-db-sync-b6wwk\" (UID: \"e43dc009-537b-4a2c-b481-e42f41eec38d\") " pod="openstack/keystone-db-sync-b6wwk" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.167302 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-65wms" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.223981 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chxft\" (UniqueName: \"kubernetes.io/projected/229ac05d-35f6-4775-8298-c61aa1378406-kube-api-access-chxft\") pod \"neutron-a32f-account-create-h2s4w\" (UID: \"229ac05d-35f6-4775-8298-c61aa1378406\") " pod="openstack/neutron-a32f-account-create-h2s4w" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.224110 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/229ac05d-35f6-4775-8298-c61aa1378406-operator-scripts\") pod \"neutron-a32f-account-create-h2s4w\" (UID: \"229ac05d-35f6-4775-8298-c61aa1378406\") " pod="openstack/neutron-a32f-account-create-h2s4w" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.224872 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/229ac05d-35f6-4775-8298-c61aa1378406-operator-scripts\") pod \"neutron-a32f-account-create-h2s4w\" (UID: \"229ac05d-35f6-4775-8298-c61aa1378406\") " pod="openstack/neutron-a32f-account-create-h2s4w" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.237525 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-2xf6v" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.238527 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chxft\" (UniqueName: \"kubernetes.io/projected/229ac05d-35f6-4775-8298-c61aa1378406-kube-api-access-chxft\") pod \"neutron-a32f-account-create-h2s4w\" (UID: \"229ac05d-35f6-4775-8298-c61aa1378406\") " pod="openstack/neutron-a32f-account-create-h2s4w" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.299316 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-b6wwk" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.304264 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-a32f-account-create-h2s4w" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.317549 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-j4kkb"] Nov 23 06:59:15 crc kubenswrapper[4559]: W1123 06:59:15.329570 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39400620_4347_45f5_85ee_4bfed29cb5cd.slice/crio-9aa35db3c0ae2bc550e8467c4ca6ca839cc89a91a6d6c5986a627987e3b2a19d WatchSource:0}: Error finding container 9aa35db3c0ae2bc550e8467c4ca6ca839cc89a91a6d6c5986a627987e3b2a19d: Status 404 returned error can't find the container with id 9aa35db3c0ae2bc550e8467c4ca6ca839cc89a91a6d6c5986a627987e3b2a19d Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.400488 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-f4d5-account-create-q6dwz"] Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.489178 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-6ad9-account-create-bfsj7"] Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.564741 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-65wms"] Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.645136 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-2xf6v"] Nov 23 06:59:15 crc kubenswrapper[4559]: W1123 06:59:15.650382 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda2b9eb8d_f298_4085_9b0e_4c9c929e23c2.slice/crio-596ae189e98d3dcc45e760b9d8d8511adb20178b698debbe3ee3076974873f0f WatchSource:0}: Error finding container 596ae189e98d3dcc45e760b9d8d8511adb20178b698debbe3ee3076974873f0f: Status 404 returned error can't find the container with id 596ae189e98d3dcc45e760b9d8d8511adb20178b698debbe3ee3076974873f0f Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.742167 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-a32f-account-create-h2s4w"] Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.749749 4559 generic.go:334] "Generic (PLEG): container finished" podID="39400620-4347-45f5-85ee-4bfed29cb5cd" containerID="69208340657fbd898448ec1513e3922261bc0f607432bd0247df66b8e7d066d2" exitCode=0 Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.749854 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-j4kkb" event={"ID":"39400620-4347-45f5-85ee-4bfed29cb5cd","Type":"ContainerDied","Data":"69208340657fbd898448ec1513e3922261bc0f607432bd0247df66b8e7d066d2"} Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.749940 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-j4kkb" event={"ID":"39400620-4347-45f5-85ee-4bfed29cb5cd","Type":"ContainerStarted","Data":"9aa35db3c0ae2bc550e8467c4ca6ca839cc89a91a6d6c5986a627987e3b2a19d"} Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.752094 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-2xf6v" event={"ID":"a2b9eb8d-f298-4085-9b0e-4c9c929e23c2","Type":"ContainerStarted","Data":"596ae189e98d3dcc45e760b9d8d8511adb20178b698debbe3ee3076974873f0f"} Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.754115 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f4d5-account-create-q6dwz" event={"ID":"11264a11-4f18-4550-a335-c0fd08428786","Type":"ContainerStarted","Data":"8ba9f205a0ae7c46a2c64cf1fa99ff0dbb88b0dd9df79a738949e5d13c872fd7"} Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.754145 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f4d5-account-create-q6dwz" event={"ID":"11264a11-4f18-4550-a335-c0fd08428786","Type":"ContainerStarted","Data":"a7128316d21dcc4e59aa35d51d3ac0d37a027a9957cfc682a3a284dcbee2025a"} Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.756328 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6ad9-account-create-bfsj7" event={"ID":"c6600577-55b0-439d-85c2-dab41fac1775","Type":"ContainerStarted","Data":"006eea70d80fe9a828e3cf95c539095e2476f9d9af8cf91aaf392f78bea994e5"} Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.756358 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6ad9-account-create-bfsj7" event={"ID":"c6600577-55b0-439d-85c2-dab41fac1775","Type":"ContainerStarted","Data":"e101f10f3adb5aaf8424fc8992912fd0f44733cd48cd8a324bdbe1d592650e41"} Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.759086 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-65wms" event={"ID":"d08d731c-9663-4eb7-ac6d-8f500dc4b294","Type":"ContainerStarted","Data":"e2961ad63ff176f2f31a2c962fb7fbcdfa68d43edcc56527f0a2400a3e0ee978"} Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.759113 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-65wms" event={"ID":"d08d731c-9663-4eb7-ac6d-8f500dc4b294","Type":"ContainerStarted","Data":"27d4938fff75a6d13b05219c40cac1dc8f989d8f5dd88df306109a5c63d61389"} Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.818979 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-6ad9-account-create-bfsj7" podStartSLOduration=1.818959076 podStartE2EDuration="1.818959076s" podCreationTimestamp="2025-11-23 06:59:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:59:15.807826478 +0000 UTC m=+857.829812092" watchObservedRunningTime="2025-11-23 06:59:15.818959076 +0000 UTC m=+857.840944690" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.822547 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-65wms" podStartSLOduration=1.8225401909999999 podStartE2EDuration="1.822540191s" podCreationTimestamp="2025-11-23 06:59:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:59:15.786397136 +0000 UTC m=+857.808382750" watchObservedRunningTime="2025-11-23 06:59:15.822540191 +0000 UTC m=+857.844525805" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.826363 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-f4d5-account-create-q6dwz" podStartSLOduration=1.8263475009999999 podStartE2EDuration="1.826347501s" podCreationTimestamp="2025-11-23 06:59:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:59:15.823045071 +0000 UTC m=+857.845030684" watchObservedRunningTime="2025-11-23 06:59:15.826347501 +0000 UTC m=+857.848333115" Nov 23 06:59:15 crc kubenswrapper[4559]: I1123 06:59:15.838824 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-b6wwk"] Nov 23 06:59:15 crc kubenswrapper[4559]: W1123 06:59:15.844339 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode43dc009_537b_4a2c_b481_e42f41eec38d.slice/crio-0f21a7d1727aa7eb5889a0fbed83f21404261fad2becbab20dc6892ddd523da6 WatchSource:0}: Error finding container 0f21a7d1727aa7eb5889a0fbed83f21404261fad2becbab20dc6892ddd523da6: Status 404 returned error can't find the container with id 0f21a7d1727aa7eb5889a0fbed83f21404261fad2becbab20dc6892ddd523da6 Nov 23 06:59:16 crc kubenswrapper[4559]: I1123 06:59:16.771611 4559 generic.go:334] "Generic (PLEG): container finished" podID="d08d731c-9663-4eb7-ac6d-8f500dc4b294" containerID="e2961ad63ff176f2f31a2c962fb7fbcdfa68d43edcc56527f0a2400a3e0ee978" exitCode=0 Nov 23 06:59:16 crc kubenswrapper[4559]: I1123 06:59:16.771720 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-65wms" event={"ID":"d08d731c-9663-4eb7-ac6d-8f500dc4b294","Type":"ContainerDied","Data":"e2961ad63ff176f2f31a2c962fb7fbcdfa68d43edcc56527f0a2400a3e0ee978"} Nov 23 06:59:16 crc kubenswrapper[4559]: I1123 06:59:16.773535 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-b6wwk" event={"ID":"e43dc009-537b-4a2c-b481-e42f41eec38d","Type":"ContainerStarted","Data":"0f21a7d1727aa7eb5889a0fbed83f21404261fad2becbab20dc6892ddd523da6"} Nov 23 06:59:16 crc kubenswrapper[4559]: I1123 06:59:16.775280 4559 generic.go:334] "Generic (PLEG): container finished" podID="a2b9eb8d-f298-4085-9b0e-4c9c929e23c2" containerID="61609ef0c1efb69623deb3e9eb896c040656881ea7d6aae5d9c4908101dd6605" exitCode=0 Nov 23 06:59:16 crc kubenswrapper[4559]: I1123 06:59:16.775346 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-2xf6v" event={"ID":"a2b9eb8d-f298-4085-9b0e-4c9c929e23c2","Type":"ContainerDied","Data":"61609ef0c1efb69623deb3e9eb896c040656881ea7d6aae5d9c4908101dd6605"} Nov 23 06:59:16 crc kubenswrapper[4559]: I1123 06:59:16.776572 4559 generic.go:334] "Generic (PLEG): container finished" podID="11264a11-4f18-4550-a335-c0fd08428786" containerID="8ba9f205a0ae7c46a2c64cf1fa99ff0dbb88b0dd9df79a738949e5d13c872fd7" exitCode=0 Nov 23 06:59:16 crc kubenswrapper[4559]: I1123 06:59:16.776614 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f4d5-account-create-q6dwz" event={"ID":"11264a11-4f18-4550-a335-c0fd08428786","Type":"ContainerDied","Data":"8ba9f205a0ae7c46a2c64cf1fa99ff0dbb88b0dd9df79a738949e5d13c872fd7"} Nov 23 06:59:16 crc kubenswrapper[4559]: I1123 06:59:16.778443 4559 generic.go:334] "Generic (PLEG): container finished" podID="c6600577-55b0-439d-85c2-dab41fac1775" containerID="006eea70d80fe9a828e3cf95c539095e2476f9d9af8cf91aaf392f78bea994e5" exitCode=0 Nov 23 06:59:16 crc kubenswrapper[4559]: I1123 06:59:16.778482 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6ad9-account-create-bfsj7" event={"ID":"c6600577-55b0-439d-85c2-dab41fac1775","Type":"ContainerDied","Data":"006eea70d80fe9a828e3cf95c539095e2476f9d9af8cf91aaf392f78bea994e5"} Nov 23 06:59:16 crc kubenswrapper[4559]: I1123 06:59:16.779890 4559 generic.go:334] "Generic (PLEG): container finished" podID="229ac05d-35f6-4775-8298-c61aa1378406" containerID="0631320823f51f57cba78a57d4205017d5ce7509f10c6009526c665185df09f3" exitCode=0 Nov 23 06:59:16 crc kubenswrapper[4559]: I1123 06:59:16.780057 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-a32f-account-create-h2s4w" event={"ID":"229ac05d-35f6-4775-8298-c61aa1378406","Type":"ContainerDied","Data":"0631320823f51f57cba78a57d4205017d5ce7509f10c6009526c665185df09f3"} Nov 23 06:59:16 crc kubenswrapper[4559]: I1123 06:59:16.780079 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-a32f-account-create-h2s4w" event={"ID":"229ac05d-35f6-4775-8298-c61aa1378406","Type":"ContainerStarted","Data":"2dc2c916da8167e6d366371d076ef3da979bdc3c441800b2b9ec3ec070102107"} Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.109316 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-j4kkb" Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.268149 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fstb\" (UniqueName: \"kubernetes.io/projected/39400620-4347-45f5-85ee-4bfed29cb5cd-kube-api-access-5fstb\") pod \"39400620-4347-45f5-85ee-4bfed29cb5cd\" (UID: \"39400620-4347-45f5-85ee-4bfed29cb5cd\") " Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.268372 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/39400620-4347-45f5-85ee-4bfed29cb5cd-operator-scripts\") pod \"39400620-4347-45f5-85ee-4bfed29cb5cd\" (UID: \"39400620-4347-45f5-85ee-4bfed29cb5cd\") " Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.269575 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39400620-4347-45f5-85ee-4bfed29cb5cd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "39400620-4347-45f5-85ee-4bfed29cb5cd" (UID: "39400620-4347-45f5-85ee-4bfed29cb5cd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.277539 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39400620-4347-45f5-85ee-4bfed29cb5cd-kube-api-access-5fstb" (OuterVolumeSpecName: "kube-api-access-5fstb") pod "39400620-4347-45f5-85ee-4bfed29cb5cd" (UID: "39400620-4347-45f5-85ee-4bfed29cb5cd"). InnerVolumeSpecName "kube-api-access-5fstb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.284916 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-648ff47655-pcz6l" Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.342155 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7dfd8c6765-hrs74"] Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.342417 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7dfd8c6765-hrs74" podUID="35d05808-5864-4a8f-ab67-e4627d8c8770" containerName="dnsmasq-dns" containerID="cri-o://991c1fdd9ab230c51a73e633f16b0abb7617d383c7887c06e344965a25700f6a" gracePeriod=10 Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.374873 4559 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/39400620-4347-45f5-85ee-4bfed29cb5cd-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.374913 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fstb\" (UniqueName: \"kubernetes.io/projected/39400620-4347-45f5-85ee-4bfed29cb5cd-kube-api-access-5fstb\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.761908 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7dfd8c6765-hrs74" Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.781324 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35d05808-5864-4a8f-ab67-e4627d8c8770-config\") pod \"35d05808-5864-4a8f-ab67-e4627d8c8770\" (UID: \"35d05808-5864-4a8f-ab67-e4627d8c8770\") " Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.781376 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35d05808-5864-4a8f-ab67-e4627d8c8770-dns-svc\") pod \"35d05808-5864-4a8f-ab67-e4627d8c8770\" (UID: \"35d05808-5864-4a8f-ab67-e4627d8c8770\") " Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.782051 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35d05808-5864-4a8f-ab67-e4627d8c8770-ovsdbserver-sb\") pod \"35d05808-5864-4a8f-ab67-e4627d8c8770\" (UID: \"35d05808-5864-4a8f-ab67-e4627d8c8770\") " Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.782213 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ql2d\" (UniqueName: \"kubernetes.io/projected/35d05808-5864-4a8f-ab67-e4627d8c8770-kube-api-access-5ql2d\") pod \"35d05808-5864-4a8f-ab67-e4627d8c8770\" (UID: \"35d05808-5864-4a8f-ab67-e4627d8c8770\") " Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.782260 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35d05808-5864-4a8f-ab67-e4627d8c8770-ovsdbserver-nb\") pod \"35d05808-5864-4a8f-ab67-e4627d8c8770\" (UID: \"35d05808-5864-4a8f-ab67-e4627d8c8770\") " Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.787526 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35d05808-5864-4a8f-ab67-e4627d8c8770-kube-api-access-5ql2d" (OuterVolumeSpecName: "kube-api-access-5ql2d") pod "35d05808-5864-4a8f-ab67-e4627d8c8770" (UID: "35d05808-5864-4a8f-ab67-e4627d8c8770"). InnerVolumeSpecName "kube-api-access-5ql2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.805429 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-j4kkb" Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.805822 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-j4kkb" event={"ID":"39400620-4347-45f5-85ee-4bfed29cb5cd","Type":"ContainerDied","Data":"9aa35db3c0ae2bc550e8467c4ca6ca839cc89a91a6d6c5986a627987e3b2a19d"} Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.805852 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9aa35db3c0ae2bc550e8467c4ca6ca839cc89a91a6d6c5986a627987e3b2a19d" Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.821782 4559 generic.go:334] "Generic (PLEG): container finished" podID="35d05808-5864-4a8f-ab67-e4627d8c8770" containerID="991c1fdd9ab230c51a73e633f16b0abb7617d383c7887c06e344965a25700f6a" exitCode=0 Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.822099 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7dfd8c6765-hrs74" Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.822146 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7dfd8c6765-hrs74" event={"ID":"35d05808-5864-4a8f-ab67-e4627d8c8770","Type":"ContainerDied","Data":"991c1fdd9ab230c51a73e633f16b0abb7617d383c7887c06e344965a25700f6a"} Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.822176 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7dfd8c6765-hrs74" event={"ID":"35d05808-5864-4a8f-ab67-e4627d8c8770","Type":"ContainerDied","Data":"8a9f56998f257076a74d5200ed3839b2095925a3322ad73d7a6b8996b321f900"} Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.822195 4559 scope.go:117] "RemoveContainer" containerID="991c1fdd9ab230c51a73e633f16b0abb7617d383c7887c06e344965a25700f6a" Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.828200 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35d05808-5864-4a8f-ab67-e4627d8c8770-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "35d05808-5864-4a8f-ab67-e4627d8c8770" (UID: "35d05808-5864-4a8f-ab67-e4627d8c8770"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.848227 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35d05808-5864-4a8f-ab67-e4627d8c8770-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "35d05808-5864-4a8f-ab67-e4627d8c8770" (UID: "35d05808-5864-4a8f-ab67-e4627d8c8770"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.850846 4559 scope.go:117] "RemoveContainer" containerID="2bf22d0a461b496602b5c1dd596999ec1d4b2319de0254ce2fdc3ace79b54bb8" Nov 23 06:59:17 crc kubenswrapper[4559]: E1123 06:59:17.851251 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/35d05808-5864-4a8f-ab67-e4627d8c8770-ovsdbserver-sb podName:35d05808-5864-4a8f-ab67-e4627d8c8770 nodeName:}" failed. No retries permitted until 2025-11-23 06:59:18.351184637 +0000 UTC m=+860.373170250 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ovsdbserver-sb" (UniqueName: "kubernetes.io/configmap/35d05808-5864-4a8f-ab67-e4627d8c8770-ovsdbserver-sb") pod "35d05808-5864-4a8f-ab67-e4627d8c8770" (UID: "35d05808-5864-4a8f-ab67-e4627d8c8770") : error deleting /var/lib/kubelet/pods/35d05808-5864-4a8f-ab67-e4627d8c8770/volume-subpaths: remove /var/lib/kubelet/pods/35d05808-5864-4a8f-ab67-e4627d8c8770/volume-subpaths: no such file or directory Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.851494 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35d05808-5864-4a8f-ab67-e4627d8c8770-config" (OuterVolumeSpecName: "config") pod "35d05808-5864-4a8f-ab67-e4627d8c8770" (UID: "35d05808-5864-4a8f-ab67-e4627d8c8770"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.885802 4559 scope.go:117] "RemoveContainer" containerID="991c1fdd9ab230c51a73e633f16b0abb7617d383c7887c06e344965a25700f6a" Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.887280 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35d05808-5864-4a8f-ab67-e4627d8c8770-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.887302 4559 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35d05808-5864-4a8f-ab67-e4627d8c8770-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.887312 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ql2d\" (UniqueName: \"kubernetes.io/projected/35d05808-5864-4a8f-ab67-e4627d8c8770-kube-api-access-5ql2d\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.887325 4559 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35d05808-5864-4a8f-ab67-e4627d8c8770-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:17 crc kubenswrapper[4559]: E1123 06:59:17.888181 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"991c1fdd9ab230c51a73e633f16b0abb7617d383c7887c06e344965a25700f6a\": container with ID starting with 991c1fdd9ab230c51a73e633f16b0abb7617d383c7887c06e344965a25700f6a not found: ID does not exist" containerID="991c1fdd9ab230c51a73e633f16b0abb7617d383c7887c06e344965a25700f6a" Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.888213 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"991c1fdd9ab230c51a73e633f16b0abb7617d383c7887c06e344965a25700f6a"} err="failed to get container status \"991c1fdd9ab230c51a73e633f16b0abb7617d383c7887c06e344965a25700f6a\": rpc error: code = NotFound desc = could not find container \"991c1fdd9ab230c51a73e633f16b0abb7617d383c7887c06e344965a25700f6a\": container with ID starting with 991c1fdd9ab230c51a73e633f16b0abb7617d383c7887c06e344965a25700f6a not found: ID does not exist" Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.888238 4559 scope.go:117] "RemoveContainer" containerID="2bf22d0a461b496602b5c1dd596999ec1d4b2319de0254ce2fdc3ace79b54bb8" Nov 23 06:59:17 crc kubenswrapper[4559]: E1123 06:59:17.889422 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bf22d0a461b496602b5c1dd596999ec1d4b2319de0254ce2fdc3ace79b54bb8\": container with ID starting with 2bf22d0a461b496602b5c1dd596999ec1d4b2319de0254ce2fdc3ace79b54bb8 not found: ID does not exist" containerID="2bf22d0a461b496602b5c1dd596999ec1d4b2319de0254ce2fdc3ace79b54bb8" Nov 23 06:59:17 crc kubenswrapper[4559]: I1123 06:59:17.889452 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bf22d0a461b496602b5c1dd596999ec1d4b2319de0254ce2fdc3ace79b54bb8"} err="failed to get container status \"2bf22d0a461b496602b5c1dd596999ec1d4b2319de0254ce2fdc3ace79b54bb8\": rpc error: code = NotFound desc = could not find container \"2bf22d0a461b496602b5c1dd596999ec1d4b2319de0254ce2fdc3ace79b54bb8\": container with ID starting with 2bf22d0a461b496602b5c1dd596999ec1d4b2319de0254ce2fdc3ace79b54bb8 not found: ID does not exist" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.224268 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f4d5-account-create-q6dwz" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.303845 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dt7mz\" (UniqueName: \"kubernetes.io/projected/11264a11-4f18-4550-a335-c0fd08428786-kube-api-access-dt7mz\") pod \"11264a11-4f18-4550-a335-c0fd08428786\" (UID: \"11264a11-4f18-4550-a335-c0fd08428786\") " Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.303914 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11264a11-4f18-4550-a335-c0fd08428786-operator-scripts\") pod \"11264a11-4f18-4550-a335-c0fd08428786\" (UID: \"11264a11-4f18-4550-a335-c0fd08428786\") " Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.304719 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11264a11-4f18-4550-a335-c0fd08428786-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "11264a11-4f18-4550-a335-c0fd08428786" (UID: "11264a11-4f18-4550-a335-c0fd08428786"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.310892 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11264a11-4f18-4550-a335-c0fd08428786-kube-api-access-dt7mz" (OuterVolumeSpecName: "kube-api-access-dt7mz") pod "11264a11-4f18-4550-a335-c0fd08428786" (UID: "11264a11-4f18-4550-a335-c0fd08428786"). InnerVolumeSpecName "kube-api-access-dt7mz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.319601 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-2xf6v" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.332361 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6ad9-account-create-bfsj7" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.332978 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-a32f-account-create-h2s4w" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.341768 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-65wms" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.405243 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6600577-55b0-439d-85c2-dab41fac1775-operator-scripts\") pod \"c6600577-55b0-439d-85c2-dab41fac1775\" (UID: \"c6600577-55b0-439d-85c2-dab41fac1775\") " Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.405294 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chxft\" (UniqueName: \"kubernetes.io/projected/229ac05d-35f6-4775-8298-c61aa1378406-kube-api-access-chxft\") pod \"229ac05d-35f6-4775-8298-c61aa1378406\" (UID: \"229ac05d-35f6-4775-8298-c61aa1378406\") " Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.405323 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35d05808-5864-4a8f-ab67-e4627d8c8770-ovsdbserver-sb\") pod \"35d05808-5864-4a8f-ab67-e4627d8c8770\" (UID: \"35d05808-5864-4a8f-ab67-e4627d8c8770\") " Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.405350 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d08d731c-9663-4eb7-ac6d-8f500dc4b294-operator-scripts\") pod \"d08d731c-9663-4eb7-ac6d-8f500dc4b294\" (UID: \"d08d731c-9663-4eb7-ac6d-8f500dc4b294\") " Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.405380 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/229ac05d-35f6-4775-8298-c61aa1378406-operator-scripts\") pod \"229ac05d-35f6-4775-8298-c61aa1378406\" (UID: \"229ac05d-35f6-4775-8298-c61aa1378406\") " Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.405445 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwl7d\" (UniqueName: \"kubernetes.io/projected/a2b9eb8d-f298-4085-9b0e-4c9c929e23c2-kube-api-access-rwl7d\") pod \"a2b9eb8d-f298-4085-9b0e-4c9c929e23c2\" (UID: \"a2b9eb8d-f298-4085-9b0e-4c9c929e23c2\") " Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.405462 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgwhn\" (UniqueName: \"kubernetes.io/projected/d08d731c-9663-4eb7-ac6d-8f500dc4b294-kube-api-access-fgwhn\") pod \"d08d731c-9663-4eb7-ac6d-8f500dc4b294\" (UID: \"d08d731c-9663-4eb7-ac6d-8f500dc4b294\") " Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.405486 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2b9eb8d-f298-4085-9b0e-4c9c929e23c2-operator-scripts\") pod \"a2b9eb8d-f298-4085-9b0e-4c9c929e23c2\" (UID: \"a2b9eb8d-f298-4085-9b0e-4c9c929e23c2\") " Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.405503 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcd2k\" (UniqueName: \"kubernetes.io/projected/c6600577-55b0-439d-85c2-dab41fac1775-kube-api-access-fcd2k\") pod \"c6600577-55b0-439d-85c2-dab41fac1775\" (UID: \"c6600577-55b0-439d-85c2-dab41fac1775\") " Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.405693 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dt7mz\" (UniqueName: \"kubernetes.io/projected/11264a11-4f18-4550-a335-c0fd08428786-kube-api-access-dt7mz\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.405705 4559 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11264a11-4f18-4550-a335-c0fd08428786-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.406222 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/229ac05d-35f6-4775-8298-c61aa1378406-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "229ac05d-35f6-4775-8298-c61aa1378406" (UID: "229ac05d-35f6-4775-8298-c61aa1378406"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.406579 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6600577-55b0-439d-85c2-dab41fac1775-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c6600577-55b0-439d-85c2-dab41fac1775" (UID: "c6600577-55b0-439d-85c2-dab41fac1775"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.406598 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35d05808-5864-4a8f-ab67-e4627d8c8770-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "35d05808-5864-4a8f-ab67-e4627d8c8770" (UID: "35d05808-5864-4a8f-ab67-e4627d8c8770"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.406598 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2b9eb8d-f298-4085-9b0e-4c9c929e23c2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a2b9eb8d-f298-4085-9b0e-4c9c929e23c2" (UID: "a2b9eb8d-f298-4085-9b0e-4c9c929e23c2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.406667 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d08d731c-9663-4eb7-ac6d-8f500dc4b294-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d08d731c-9663-4eb7-ac6d-8f500dc4b294" (UID: "d08d731c-9663-4eb7-ac6d-8f500dc4b294"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.408723 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6600577-55b0-439d-85c2-dab41fac1775-kube-api-access-fcd2k" (OuterVolumeSpecName: "kube-api-access-fcd2k") pod "c6600577-55b0-439d-85c2-dab41fac1775" (UID: "c6600577-55b0-439d-85c2-dab41fac1775"). InnerVolumeSpecName "kube-api-access-fcd2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.408764 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d08d731c-9663-4eb7-ac6d-8f500dc4b294-kube-api-access-fgwhn" (OuterVolumeSpecName: "kube-api-access-fgwhn") pod "d08d731c-9663-4eb7-ac6d-8f500dc4b294" (UID: "d08d731c-9663-4eb7-ac6d-8f500dc4b294"). InnerVolumeSpecName "kube-api-access-fgwhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.409108 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/229ac05d-35f6-4775-8298-c61aa1378406-kube-api-access-chxft" (OuterVolumeSpecName: "kube-api-access-chxft") pod "229ac05d-35f6-4775-8298-c61aa1378406" (UID: "229ac05d-35f6-4775-8298-c61aa1378406"). InnerVolumeSpecName "kube-api-access-chxft". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.409202 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2b9eb8d-f298-4085-9b0e-4c9c929e23c2-kube-api-access-rwl7d" (OuterVolumeSpecName: "kube-api-access-rwl7d") pod "a2b9eb8d-f298-4085-9b0e-4c9c929e23c2" (UID: "a2b9eb8d-f298-4085-9b0e-4c9c929e23c2"). InnerVolumeSpecName "kube-api-access-rwl7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.450424 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7dfd8c6765-hrs74"] Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.454359 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7dfd8c6765-hrs74"] Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.506944 4559 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6600577-55b0-439d-85c2-dab41fac1775-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.506971 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chxft\" (UniqueName: \"kubernetes.io/projected/229ac05d-35f6-4775-8298-c61aa1378406-kube-api-access-chxft\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.506980 4559 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35d05808-5864-4a8f-ab67-e4627d8c8770-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.506988 4559 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d08d731c-9663-4eb7-ac6d-8f500dc4b294-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.506998 4559 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/229ac05d-35f6-4775-8298-c61aa1378406-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.507006 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwl7d\" (UniqueName: \"kubernetes.io/projected/a2b9eb8d-f298-4085-9b0e-4c9c929e23c2-kube-api-access-rwl7d\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.507014 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgwhn\" (UniqueName: \"kubernetes.io/projected/d08d731c-9663-4eb7-ac6d-8f500dc4b294-kube-api-access-fgwhn\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.507021 4559 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2b9eb8d-f298-4085-9b0e-4c9c929e23c2-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.507029 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcd2k\" (UniqueName: \"kubernetes.io/projected/c6600577-55b0-439d-85c2-dab41fac1775-kube-api-access-fcd2k\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.836854 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-65wms" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.836837 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-65wms" event={"ID":"d08d731c-9663-4eb7-ac6d-8f500dc4b294","Type":"ContainerDied","Data":"27d4938fff75a6d13b05219c40cac1dc8f989d8f5dd88df306109a5c63d61389"} Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.836991 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27d4938fff75a6d13b05219c40cac1dc8f989d8f5dd88df306109a5c63d61389" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.838100 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-2xf6v" event={"ID":"a2b9eb8d-f298-4085-9b0e-4c9c929e23c2","Type":"ContainerDied","Data":"596ae189e98d3dcc45e760b9d8d8511adb20178b698debbe3ee3076974873f0f"} Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.838149 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="596ae189e98d3dcc45e760b9d8d8511adb20178b698debbe3ee3076974873f0f" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.838167 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-2xf6v" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.839326 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f4d5-account-create-q6dwz" event={"ID":"11264a11-4f18-4550-a335-c0fd08428786","Type":"ContainerDied","Data":"a7128316d21dcc4e59aa35d51d3ac0d37a027a9957cfc682a3a284dcbee2025a"} Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.839337 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f4d5-account-create-q6dwz" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.839353 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7128316d21dcc4e59aa35d51d3ac0d37a027a9957cfc682a3a284dcbee2025a" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.840859 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6ad9-account-create-bfsj7" event={"ID":"c6600577-55b0-439d-85c2-dab41fac1775","Type":"ContainerDied","Data":"e101f10f3adb5aaf8424fc8992912fd0f44733cd48cd8a324bdbe1d592650e41"} Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.840878 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6ad9-account-create-bfsj7" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.840884 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e101f10f3adb5aaf8424fc8992912fd0f44733cd48cd8a324bdbe1d592650e41" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.842088 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-a32f-account-create-h2s4w" event={"ID":"229ac05d-35f6-4775-8298-c61aa1378406","Type":"ContainerDied","Data":"2dc2c916da8167e6d366371d076ef3da979bdc3c441800b2b9ec3ec070102107"} Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.842125 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2dc2c916da8167e6d366371d076ef3da979bdc3c441800b2b9ec3ec070102107" Nov 23 06:59:18 crc kubenswrapper[4559]: I1123 06:59:18.842096 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-a32f-account-create-h2s4w" Nov 23 06:59:20 crc kubenswrapper[4559]: I1123 06:59:20.305587 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35d05808-5864-4a8f-ab67-e4627d8c8770" path="/var/lib/kubelet/pods/35d05808-5864-4a8f-ab67-e4627d8c8770/volumes" Nov 23 06:59:21 crc kubenswrapper[4559]: I1123 06:59:21.864154 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-b6wwk" event={"ID":"e43dc009-537b-4a2c-b481-e42f41eec38d","Type":"ContainerStarted","Data":"4b7bde1d24163aa6ba263b550d46b5c17d85dad13c90a85a22611065fe68b406"} Nov 23 06:59:21 crc kubenswrapper[4559]: I1123 06:59:21.881473 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-b6wwk" podStartSLOduration=3.014807464 podStartE2EDuration="7.881455334s" podCreationTimestamp="2025-11-23 06:59:14 +0000 UTC" firstStartedPulling="2025-11-23 06:59:15.847282825 +0000 UTC m=+857.869268439" lastFinishedPulling="2025-11-23 06:59:20.713930694 +0000 UTC m=+862.735916309" observedRunningTime="2025-11-23 06:59:21.876591566 +0000 UTC m=+863.898577180" watchObservedRunningTime="2025-11-23 06:59:21.881455334 +0000 UTC m=+863.903440948" Nov 23 06:59:22 crc kubenswrapper[4559]: I1123 06:59:22.872792 4559 generic.go:334] "Generic (PLEG): container finished" podID="e43dc009-537b-4a2c-b481-e42f41eec38d" containerID="4b7bde1d24163aa6ba263b550d46b5c17d85dad13c90a85a22611065fe68b406" exitCode=0 Nov 23 06:59:22 crc kubenswrapper[4559]: I1123 06:59:22.872837 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-b6wwk" event={"ID":"e43dc009-537b-4a2c-b481-e42f41eec38d","Type":"ContainerDied","Data":"4b7bde1d24163aa6ba263b550d46b5c17d85dad13c90a85a22611065fe68b406"} Nov 23 06:59:24 crc kubenswrapper[4559]: I1123 06:59:24.132712 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-b6wwk" Nov 23 06:59:24 crc kubenswrapper[4559]: I1123 06:59:24.196289 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8cv2\" (UniqueName: \"kubernetes.io/projected/e43dc009-537b-4a2c-b481-e42f41eec38d-kube-api-access-f8cv2\") pod \"e43dc009-537b-4a2c-b481-e42f41eec38d\" (UID: \"e43dc009-537b-4a2c-b481-e42f41eec38d\") " Nov 23 06:59:24 crc kubenswrapper[4559]: I1123 06:59:24.196364 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e43dc009-537b-4a2c-b481-e42f41eec38d-combined-ca-bundle\") pod \"e43dc009-537b-4a2c-b481-e42f41eec38d\" (UID: \"e43dc009-537b-4a2c-b481-e42f41eec38d\") " Nov 23 06:59:24 crc kubenswrapper[4559]: I1123 06:59:24.196441 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e43dc009-537b-4a2c-b481-e42f41eec38d-config-data\") pod \"e43dc009-537b-4a2c-b481-e42f41eec38d\" (UID: \"e43dc009-537b-4a2c-b481-e42f41eec38d\") " Nov 23 06:59:24 crc kubenswrapper[4559]: I1123 06:59:24.201181 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e43dc009-537b-4a2c-b481-e42f41eec38d-kube-api-access-f8cv2" (OuterVolumeSpecName: "kube-api-access-f8cv2") pod "e43dc009-537b-4a2c-b481-e42f41eec38d" (UID: "e43dc009-537b-4a2c-b481-e42f41eec38d"). InnerVolumeSpecName "kube-api-access-f8cv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:59:24 crc kubenswrapper[4559]: I1123 06:59:24.215934 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e43dc009-537b-4a2c-b481-e42f41eec38d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e43dc009-537b-4a2c-b481-e42f41eec38d" (UID: "e43dc009-537b-4a2c-b481-e42f41eec38d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:24 crc kubenswrapper[4559]: I1123 06:59:24.229941 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e43dc009-537b-4a2c-b481-e42f41eec38d-config-data" (OuterVolumeSpecName: "config-data") pod "e43dc009-537b-4a2c-b481-e42f41eec38d" (UID: "e43dc009-537b-4a2c-b481-e42f41eec38d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:24 crc kubenswrapper[4559]: I1123 06:59:24.297605 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e43dc009-537b-4a2c-b481-e42f41eec38d-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:24 crc kubenswrapper[4559]: I1123 06:59:24.297626 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8cv2\" (UniqueName: \"kubernetes.io/projected/e43dc009-537b-4a2c-b481-e42f41eec38d-kube-api-access-f8cv2\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:24 crc kubenswrapper[4559]: I1123 06:59:24.297637 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e43dc009-537b-4a2c-b481-e42f41eec38d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:24 crc kubenswrapper[4559]: I1123 06:59:24.887394 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-b6wwk" event={"ID":"e43dc009-537b-4a2c-b481-e42f41eec38d","Type":"ContainerDied","Data":"0f21a7d1727aa7eb5889a0fbed83f21404261fad2becbab20dc6892ddd523da6"} Nov 23 06:59:24 crc kubenswrapper[4559]: I1123 06:59:24.887449 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-b6wwk" Nov 23 06:59:24 crc kubenswrapper[4559]: I1123 06:59:24.887667 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f21a7d1727aa7eb5889a0fbed83f21404261fad2becbab20dc6892ddd523da6" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.089910 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b54fd9f79-pp688"] Nov 23 06:59:25 crc kubenswrapper[4559]: E1123 06:59:25.090214 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35d05808-5864-4a8f-ab67-e4627d8c8770" containerName="init" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.090227 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="35d05808-5864-4a8f-ab67-e4627d8c8770" containerName="init" Nov 23 06:59:25 crc kubenswrapper[4559]: E1123 06:59:25.090239 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6600577-55b0-439d-85c2-dab41fac1775" containerName="mariadb-account-create" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.090245 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6600577-55b0-439d-85c2-dab41fac1775" containerName="mariadb-account-create" Nov 23 06:59:25 crc kubenswrapper[4559]: E1123 06:59:25.090255 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d08d731c-9663-4eb7-ac6d-8f500dc4b294" containerName="mariadb-database-create" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.090261 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="d08d731c-9663-4eb7-ac6d-8f500dc4b294" containerName="mariadb-database-create" Nov 23 06:59:25 crc kubenswrapper[4559]: E1123 06:59:25.090271 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e43dc009-537b-4a2c-b481-e42f41eec38d" containerName="keystone-db-sync" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.090278 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="e43dc009-537b-4a2c-b481-e42f41eec38d" containerName="keystone-db-sync" Nov 23 06:59:25 crc kubenswrapper[4559]: E1123 06:59:25.090285 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35d05808-5864-4a8f-ab67-e4627d8c8770" containerName="dnsmasq-dns" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.090291 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="35d05808-5864-4a8f-ab67-e4627d8c8770" containerName="dnsmasq-dns" Nov 23 06:59:25 crc kubenswrapper[4559]: E1123 06:59:25.090303 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2b9eb8d-f298-4085-9b0e-4c9c929e23c2" containerName="mariadb-database-create" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.090309 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2b9eb8d-f298-4085-9b0e-4c9c929e23c2" containerName="mariadb-database-create" Nov 23 06:59:25 crc kubenswrapper[4559]: E1123 06:59:25.090318 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11264a11-4f18-4550-a335-c0fd08428786" containerName="mariadb-account-create" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.090329 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="11264a11-4f18-4550-a335-c0fd08428786" containerName="mariadb-account-create" Nov 23 06:59:25 crc kubenswrapper[4559]: E1123 06:59:25.090340 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="229ac05d-35f6-4775-8298-c61aa1378406" containerName="mariadb-account-create" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.090346 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="229ac05d-35f6-4775-8298-c61aa1378406" containerName="mariadb-account-create" Nov 23 06:59:25 crc kubenswrapper[4559]: E1123 06:59:25.090354 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39400620-4347-45f5-85ee-4bfed29cb5cd" containerName="mariadb-database-create" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.090359 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="39400620-4347-45f5-85ee-4bfed29cb5cd" containerName="mariadb-database-create" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.090521 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="229ac05d-35f6-4775-8298-c61aa1378406" containerName="mariadb-account-create" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.090532 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="e43dc009-537b-4a2c-b481-e42f41eec38d" containerName="keystone-db-sync" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.090544 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6600577-55b0-439d-85c2-dab41fac1775" containerName="mariadb-account-create" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.090554 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="11264a11-4f18-4550-a335-c0fd08428786" containerName="mariadb-account-create" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.090561 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2b9eb8d-f298-4085-9b0e-4c9c929e23c2" containerName="mariadb-database-create" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.090567 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="39400620-4347-45f5-85ee-4bfed29cb5cd" containerName="mariadb-database-create" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.090581 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="35d05808-5864-4a8f-ab67-e4627d8c8770" containerName="dnsmasq-dns" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.090592 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="d08d731c-9663-4eb7-ac6d-8f500dc4b294" containerName="mariadb-database-create" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.091438 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b54fd9f79-pp688" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.110414 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b54fd9f79-pp688"] Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.152602 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-nm8sn"] Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.154204 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nm8sn" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.158284 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.158317 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.158799 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.159001 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xmhpd" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.159149 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.162989 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nm8sn"] Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.208941 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn4rt\" (UniqueName: \"kubernetes.io/projected/aa3c4b78-d545-4bf5-95cd-33ef6670e895-kube-api-access-qn4rt\") pod \"dnsmasq-dns-5b54fd9f79-pp688\" (UID: \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\") " pod="openstack/dnsmasq-dns-5b54fd9f79-pp688" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.208991 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-ovsdbserver-sb\") pod \"dnsmasq-dns-5b54fd9f79-pp688\" (UID: \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\") " pod="openstack/dnsmasq-dns-5b54fd9f79-pp688" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.209010 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-config\") pod \"dnsmasq-dns-5b54fd9f79-pp688\" (UID: \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\") " pod="openstack/dnsmasq-dns-5b54fd9f79-pp688" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.209779 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-dns-swift-storage-0\") pod \"dnsmasq-dns-5b54fd9f79-pp688\" (UID: \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\") " pod="openstack/dnsmasq-dns-5b54fd9f79-pp688" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.210494 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-dns-svc\") pod \"dnsmasq-dns-5b54fd9f79-pp688\" (UID: \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\") " pod="openstack/dnsmasq-dns-5b54fd9f79-pp688" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.210666 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-ovsdbserver-nb\") pod \"dnsmasq-dns-5b54fd9f79-pp688\" (UID: \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\") " pod="openstack/dnsmasq-dns-5b54fd9f79-pp688" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.260301 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.262006 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.263955 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.264145 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.279709 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.312792 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-fernet-keys\") pod \"keystone-bootstrap-nm8sn\" (UID: \"18ba7210-7560-48f6-b164-780a6c84ef5f\") " pod="openstack/keystone-bootstrap-nm8sn" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.313250 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzs77\" (UniqueName: \"kubernetes.io/projected/18ba7210-7560-48f6-b164-780a6c84ef5f-kube-api-access-vzs77\") pod \"keystone-bootstrap-nm8sn\" (UID: \"18ba7210-7560-48f6-b164-780a6c84ef5f\") " pod="openstack/keystone-bootstrap-nm8sn" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.313297 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-dns-svc\") pod \"dnsmasq-dns-5b54fd9f79-pp688\" (UID: \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\") " pod="openstack/dnsmasq-dns-5b54fd9f79-pp688" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.313352 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-credential-keys\") pod \"keystone-bootstrap-nm8sn\" (UID: \"18ba7210-7560-48f6-b164-780a6c84ef5f\") " pod="openstack/keystone-bootstrap-nm8sn" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.313392 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-ovsdbserver-nb\") pod \"dnsmasq-dns-5b54fd9f79-pp688\" (UID: \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\") " pod="openstack/dnsmasq-dns-5b54fd9f79-pp688" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.313407 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-combined-ca-bundle\") pod \"keystone-bootstrap-nm8sn\" (UID: \"18ba7210-7560-48f6-b164-780a6c84ef5f\") " pod="openstack/keystone-bootstrap-nm8sn" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.313509 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-config-data\") pod \"keystone-bootstrap-nm8sn\" (UID: \"18ba7210-7560-48f6-b164-780a6c84ef5f\") " pod="openstack/keystone-bootstrap-nm8sn" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.313545 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-scripts\") pod \"keystone-bootstrap-nm8sn\" (UID: \"18ba7210-7560-48f6-b164-780a6c84ef5f\") " pod="openstack/keystone-bootstrap-nm8sn" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.313563 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn4rt\" (UniqueName: \"kubernetes.io/projected/aa3c4b78-d545-4bf5-95cd-33ef6670e895-kube-api-access-qn4rt\") pod \"dnsmasq-dns-5b54fd9f79-pp688\" (UID: \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\") " pod="openstack/dnsmasq-dns-5b54fd9f79-pp688" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.313618 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-ovsdbserver-sb\") pod \"dnsmasq-dns-5b54fd9f79-pp688\" (UID: \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\") " pod="openstack/dnsmasq-dns-5b54fd9f79-pp688" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.314192 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-dns-svc\") pod \"dnsmasq-dns-5b54fd9f79-pp688\" (UID: \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\") " pod="openstack/dnsmasq-dns-5b54fd9f79-pp688" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.314196 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-config\") pod \"dnsmasq-dns-5b54fd9f79-pp688\" (UID: \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\") " pod="openstack/dnsmasq-dns-5b54fd9f79-pp688" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.314251 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-dns-swift-storage-0\") pod \"dnsmasq-dns-5b54fd9f79-pp688\" (UID: \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\") " pod="openstack/dnsmasq-dns-5b54fd9f79-pp688" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.314320 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-ovsdbserver-sb\") pod \"dnsmasq-dns-5b54fd9f79-pp688\" (UID: \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\") " pod="openstack/dnsmasq-dns-5b54fd9f79-pp688" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.314632 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-ovsdbserver-nb\") pod \"dnsmasq-dns-5b54fd9f79-pp688\" (UID: \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\") " pod="openstack/dnsmasq-dns-5b54fd9f79-pp688" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.314915 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-dns-swift-storage-0\") pod \"dnsmasq-dns-5b54fd9f79-pp688\" (UID: \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\") " pod="openstack/dnsmasq-dns-5b54fd9f79-pp688" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.315399 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-config\") pod \"dnsmasq-dns-5b54fd9f79-pp688\" (UID: \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\") " pod="openstack/dnsmasq-dns-5b54fd9f79-pp688" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.338388 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn4rt\" (UniqueName: \"kubernetes.io/projected/aa3c4b78-d545-4bf5-95cd-33ef6670e895-kube-api-access-qn4rt\") pod \"dnsmasq-dns-5b54fd9f79-pp688\" (UID: \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\") " pod="openstack/dnsmasq-dns-5b54fd9f79-pp688" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.343629 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-bxpjp"] Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.353799 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-bxpjp"] Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.353972 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-bxpjp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.356577 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.356830 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-7j26t" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.356986 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.409892 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b54fd9f79-pp688" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.415297 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8ae1721-4ded-4e7d-9091-88ac908d0554-config-data\") pod \"cinder-db-sync-bxpjp\" (UID: \"c8ae1721-4ded-4e7d-9091-88ac908d0554\") " pod="openstack/cinder-db-sync-bxpjp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.415380 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-config-data\") pod \"keystone-bootstrap-nm8sn\" (UID: \"18ba7210-7560-48f6-b164-780a6c84ef5f\") " pod="openstack/keystone-bootstrap-nm8sn" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.415434 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-scripts\") pod \"keystone-bootstrap-nm8sn\" (UID: \"18ba7210-7560-48f6-b164-780a6c84ef5f\") " pod="openstack/keystone-bootstrap-nm8sn" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.415470 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4163bfe8-0de7-4490-b825-55d0f0343787-scripts\") pod \"ceilometer-0\" (UID: \"4163bfe8-0de7-4490-b825-55d0f0343787\") " pod="openstack/ceilometer-0" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.415513 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c8ae1721-4ded-4e7d-9091-88ac908d0554-etc-machine-id\") pod \"cinder-db-sync-bxpjp\" (UID: \"c8ae1721-4ded-4e7d-9091-88ac908d0554\") " pod="openstack/cinder-db-sync-bxpjp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.415536 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxzxp\" (UniqueName: \"kubernetes.io/projected/c8ae1721-4ded-4e7d-9091-88ac908d0554-kube-api-access-pxzxp\") pod \"cinder-db-sync-bxpjp\" (UID: \"c8ae1721-4ded-4e7d-9091-88ac908d0554\") " pod="openstack/cinder-db-sync-bxpjp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.415574 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gf5bp\" (UniqueName: \"kubernetes.io/projected/4163bfe8-0de7-4490-b825-55d0f0343787-kube-api-access-gf5bp\") pod \"ceilometer-0\" (UID: \"4163bfe8-0de7-4490-b825-55d0f0343787\") " pod="openstack/ceilometer-0" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.415602 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4163bfe8-0de7-4490-b825-55d0f0343787-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4163bfe8-0de7-4490-b825-55d0f0343787\") " pod="openstack/ceilometer-0" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.415632 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4163bfe8-0de7-4490-b825-55d0f0343787-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4163bfe8-0de7-4490-b825-55d0f0343787\") " pod="openstack/ceilometer-0" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.415670 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8ae1721-4ded-4e7d-9091-88ac908d0554-combined-ca-bundle\") pod \"cinder-db-sync-bxpjp\" (UID: \"c8ae1721-4ded-4e7d-9091-88ac908d0554\") " pod="openstack/cinder-db-sync-bxpjp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.415696 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-fernet-keys\") pod \"keystone-bootstrap-nm8sn\" (UID: \"18ba7210-7560-48f6-b164-780a6c84ef5f\") " pod="openstack/keystone-bootstrap-nm8sn" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.415715 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8ae1721-4ded-4e7d-9091-88ac908d0554-scripts\") pod \"cinder-db-sync-bxpjp\" (UID: \"c8ae1721-4ded-4e7d-9091-88ac908d0554\") " pod="openstack/cinder-db-sync-bxpjp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.415748 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4163bfe8-0de7-4490-b825-55d0f0343787-run-httpd\") pod \"ceilometer-0\" (UID: \"4163bfe8-0de7-4490-b825-55d0f0343787\") " pod="openstack/ceilometer-0" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.415776 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4163bfe8-0de7-4490-b825-55d0f0343787-config-data\") pod \"ceilometer-0\" (UID: \"4163bfe8-0de7-4490-b825-55d0f0343787\") " pod="openstack/ceilometer-0" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.415918 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzs77\" (UniqueName: \"kubernetes.io/projected/18ba7210-7560-48f6-b164-780a6c84ef5f-kube-api-access-vzs77\") pod \"keystone-bootstrap-nm8sn\" (UID: \"18ba7210-7560-48f6-b164-780a6c84ef5f\") " pod="openstack/keystone-bootstrap-nm8sn" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.415936 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c8ae1721-4ded-4e7d-9091-88ac908d0554-db-sync-config-data\") pod \"cinder-db-sync-bxpjp\" (UID: \"c8ae1721-4ded-4e7d-9091-88ac908d0554\") " pod="openstack/cinder-db-sync-bxpjp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.415988 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-credential-keys\") pod \"keystone-bootstrap-nm8sn\" (UID: \"18ba7210-7560-48f6-b164-780a6c84ef5f\") " pod="openstack/keystone-bootstrap-nm8sn" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.416009 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4163bfe8-0de7-4490-b825-55d0f0343787-log-httpd\") pod \"ceilometer-0\" (UID: \"4163bfe8-0de7-4490-b825-55d0f0343787\") " pod="openstack/ceilometer-0" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.416033 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-combined-ca-bundle\") pod \"keystone-bootstrap-nm8sn\" (UID: \"18ba7210-7560-48f6-b164-780a6c84ef5f\") " pod="openstack/keystone-bootstrap-nm8sn" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.420530 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-config-data\") pod \"keystone-bootstrap-nm8sn\" (UID: \"18ba7210-7560-48f6-b164-780a6c84ef5f\") " pod="openstack/keystone-bootstrap-nm8sn" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.428074 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-combined-ca-bundle\") pod \"keystone-bootstrap-nm8sn\" (UID: \"18ba7210-7560-48f6-b164-780a6c84ef5f\") " pod="openstack/keystone-bootstrap-nm8sn" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.428140 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-scripts\") pod \"keystone-bootstrap-nm8sn\" (UID: \"18ba7210-7560-48f6-b164-780a6c84ef5f\") " pod="openstack/keystone-bootstrap-nm8sn" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.433696 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-credential-keys\") pod \"keystone-bootstrap-nm8sn\" (UID: \"18ba7210-7560-48f6-b164-780a6c84ef5f\") " pod="openstack/keystone-bootstrap-nm8sn" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.438698 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-k6zjg"] Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.442699 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-fernet-keys\") pod \"keystone-bootstrap-nm8sn\" (UID: \"18ba7210-7560-48f6-b164-780a6c84ef5f\") " pod="openstack/keystone-bootstrap-nm8sn" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.447325 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-k6zjg" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.449726 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b54fd9f79-pp688"] Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.451936 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.452137 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-lztfj" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.468750 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-qgn9k"] Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.469857 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-qgn9k" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.475495 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-bqc2b" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.475662 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.476126 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.488236 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-k6zjg"] Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.492884 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzs77\" (UniqueName: \"kubernetes.io/projected/18ba7210-7560-48f6-b164-780a6c84ef5f-kube-api-access-vzs77\") pod \"keystone-bootstrap-nm8sn\" (UID: \"18ba7210-7560-48f6-b164-780a6c84ef5f\") " pod="openstack/keystone-bootstrap-nm8sn" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.496716 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-qgn9k"] Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.506768 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-mvljp"] Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.507947 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-mvljp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.509929 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.510118 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-rdr8l" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.510337 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.510861 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d78ff46f5-kdzjr"] Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.512150 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.517750 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c8ae1721-4ded-4e7d-9091-88ac908d0554-etc-machine-id\") pod \"cinder-db-sync-bxpjp\" (UID: \"c8ae1721-4ded-4e7d-9091-88ac908d0554\") " pod="openstack/cinder-db-sync-bxpjp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.517792 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxzxp\" (UniqueName: \"kubernetes.io/projected/c8ae1721-4ded-4e7d-9091-88ac908d0554-kube-api-access-pxzxp\") pod \"cinder-db-sync-bxpjp\" (UID: \"c8ae1721-4ded-4e7d-9091-88ac908d0554\") " pod="openstack/cinder-db-sync-bxpjp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.517821 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2xnv\" (UniqueName: \"kubernetes.io/projected/ad2af1c5-49df-4093-a38f-b1b47e2ddc1f-kube-api-access-c2xnv\") pod \"neutron-db-sync-qgn9k\" (UID: \"ad2af1c5-49df-4093-a38f-b1b47e2ddc1f\") " pod="openstack/neutron-db-sync-qgn9k" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.517843 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gf5bp\" (UniqueName: \"kubernetes.io/projected/4163bfe8-0de7-4490-b825-55d0f0343787-kube-api-access-gf5bp\") pod \"ceilometer-0\" (UID: \"4163bfe8-0de7-4490-b825-55d0f0343787\") " pod="openstack/ceilometer-0" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.517864 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4163bfe8-0de7-4490-b825-55d0f0343787-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4163bfe8-0de7-4490-b825-55d0f0343787\") " pod="openstack/ceilometer-0" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.517885 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4163bfe8-0de7-4490-b825-55d0f0343787-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4163bfe8-0de7-4490-b825-55d0f0343787\") " pod="openstack/ceilometer-0" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.517910 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-dns-svc\") pod \"dnsmasq-dns-d78ff46f5-kdzjr\" (UID: \"58898cc0-18e6-41c6-9ce2-149d081ff29b\") " pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.517925 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8ae1721-4ded-4e7d-9091-88ac908d0554-combined-ca-bundle\") pod \"cinder-db-sync-bxpjp\" (UID: \"c8ae1721-4ded-4e7d-9091-88ac908d0554\") " pod="openstack/cinder-db-sync-bxpjp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.517941 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43aed302-5f95-4e51-b845-5a9cee9d1a08-scripts\") pod \"placement-db-sync-mvljp\" (UID: \"43aed302-5f95-4e51-b845-5a9cee9d1a08\") " pod="openstack/placement-db-sync-mvljp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.517960 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8ae1721-4ded-4e7d-9091-88ac908d0554-scripts\") pod \"cinder-db-sync-bxpjp\" (UID: \"c8ae1721-4ded-4e7d-9091-88ac908d0554\") " pod="openstack/cinder-db-sync-bxpjp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.517987 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad2af1c5-49df-4093-a38f-b1b47e2ddc1f-combined-ca-bundle\") pod \"neutron-db-sync-qgn9k\" (UID: \"ad2af1c5-49df-4093-a38f-b1b47e2ddc1f\") " pod="openstack/neutron-db-sync-qgn9k" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.518002 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4163bfe8-0de7-4490-b825-55d0f0343787-run-httpd\") pod \"ceilometer-0\" (UID: \"4163bfe8-0de7-4490-b825-55d0f0343787\") " pod="openstack/ceilometer-0" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.518018 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jn9c7\" (UniqueName: \"kubernetes.io/projected/b14d2018-932c-47b6-bba9-232dff3ce320-kube-api-access-jn9c7\") pod \"barbican-db-sync-k6zjg\" (UID: \"b14d2018-932c-47b6-bba9-232dff3ce320\") " pod="openstack/barbican-db-sync-k6zjg" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.518035 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-dns-swift-storage-0\") pod \"dnsmasq-dns-d78ff46f5-kdzjr\" (UID: \"58898cc0-18e6-41c6-9ce2-149d081ff29b\") " pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.518049 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m77lh\" (UniqueName: \"kubernetes.io/projected/58898cc0-18e6-41c6-9ce2-149d081ff29b-kube-api-access-m77lh\") pod \"dnsmasq-dns-d78ff46f5-kdzjr\" (UID: \"58898cc0-18e6-41c6-9ce2-149d081ff29b\") " pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.518065 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4163bfe8-0de7-4490-b825-55d0f0343787-config-data\") pod \"ceilometer-0\" (UID: \"4163bfe8-0de7-4490-b825-55d0f0343787\") " pod="openstack/ceilometer-0" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.518083 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43aed302-5f95-4e51-b845-5a9cee9d1a08-config-data\") pod \"placement-db-sync-mvljp\" (UID: \"43aed302-5f95-4e51-b845-5a9cee9d1a08\") " pod="openstack/placement-db-sync-mvljp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.518114 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k557r\" (UniqueName: \"kubernetes.io/projected/43aed302-5f95-4e51-b845-5a9cee9d1a08-kube-api-access-k557r\") pod \"placement-db-sync-mvljp\" (UID: \"43aed302-5f95-4e51-b845-5a9cee9d1a08\") " pod="openstack/placement-db-sync-mvljp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.518129 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-config\") pod \"dnsmasq-dns-d78ff46f5-kdzjr\" (UID: \"58898cc0-18e6-41c6-9ce2-149d081ff29b\") " pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.522681 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43aed302-5f95-4e51-b845-5a9cee9d1a08-logs\") pod \"placement-db-sync-mvljp\" (UID: \"43aed302-5f95-4e51-b845-5a9cee9d1a08\") " pod="openstack/placement-db-sync-mvljp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.522757 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c8ae1721-4ded-4e7d-9091-88ac908d0554-db-sync-config-data\") pod \"cinder-db-sync-bxpjp\" (UID: \"c8ae1721-4ded-4e7d-9091-88ac908d0554\") " pod="openstack/cinder-db-sync-bxpjp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.522777 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b14d2018-932c-47b6-bba9-232dff3ce320-db-sync-config-data\") pod \"barbican-db-sync-k6zjg\" (UID: \"b14d2018-932c-47b6-bba9-232dff3ce320\") " pod="openstack/barbican-db-sync-k6zjg" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.522802 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-ovsdbserver-nb\") pod \"dnsmasq-dns-d78ff46f5-kdzjr\" (UID: \"58898cc0-18e6-41c6-9ce2-149d081ff29b\") " pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.522837 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4163bfe8-0de7-4490-b825-55d0f0343787-log-httpd\") pod \"ceilometer-0\" (UID: \"4163bfe8-0de7-4490-b825-55d0f0343787\") " pod="openstack/ceilometer-0" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.522858 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b14d2018-932c-47b6-bba9-232dff3ce320-combined-ca-bundle\") pod \"barbican-db-sync-k6zjg\" (UID: \"b14d2018-932c-47b6-bba9-232dff3ce320\") " pod="openstack/barbican-db-sync-k6zjg" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.522883 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8ae1721-4ded-4e7d-9091-88ac908d0554-config-data\") pod \"cinder-db-sync-bxpjp\" (UID: \"c8ae1721-4ded-4e7d-9091-88ac908d0554\") " pod="openstack/cinder-db-sync-bxpjp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.522918 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-ovsdbserver-sb\") pod \"dnsmasq-dns-d78ff46f5-kdzjr\" (UID: \"58898cc0-18e6-41c6-9ce2-149d081ff29b\") " pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.522947 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43aed302-5f95-4e51-b845-5a9cee9d1a08-combined-ca-bundle\") pod \"placement-db-sync-mvljp\" (UID: \"43aed302-5f95-4e51-b845-5a9cee9d1a08\") " pod="openstack/placement-db-sync-mvljp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.522977 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4163bfe8-0de7-4490-b825-55d0f0343787-scripts\") pod \"ceilometer-0\" (UID: \"4163bfe8-0de7-4490-b825-55d0f0343787\") " pod="openstack/ceilometer-0" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.522989 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ad2af1c5-49df-4093-a38f-b1b47e2ddc1f-config\") pod \"neutron-db-sync-qgn9k\" (UID: \"ad2af1c5-49df-4093-a38f-b1b47e2ddc1f\") " pod="openstack/neutron-db-sync-qgn9k" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.523053 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c8ae1721-4ded-4e7d-9091-88ac908d0554-etc-machine-id\") pod \"cinder-db-sync-bxpjp\" (UID: \"c8ae1721-4ded-4e7d-9091-88ac908d0554\") " pod="openstack/cinder-db-sync-bxpjp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.527675 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4163bfe8-0de7-4490-b825-55d0f0343787-log-httpd\") pod \"ceilometer-0\" (UID: \"4163bfe8-0de7-4490-b825-55d0f0343787\") " pod="openstack/ceilometer-0" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.530087 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4163bfe8-0de7-4490-b825-55d0f0343787-run-httpd\") pod \"ceilometer-0\" (UID: \"4163bfe8-0de7-4490-b825-55d0f0343787\") " pod="openstack/ceilometer-0" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.539338 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c8ae1721-4ded-4e7d-9091-88ac908d0554-db-sync-config-data\") pod \"cinder-db-sync-bxpjp\" (UID: \"c8ae1721-4ded-4e7d-9091-88ac908d0554\") " pod="openstack/cinder-db-sync-bxpjp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.540447 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8ae1721-4ded-4e7d-9091-88ac908d0554-scripts\") pod \"cinder-db-sync-bxpjp\" (UID: \"c8ae1721-4ded-4e7d-9091-88ac908d0554\") " pod="openstack/cinder-db-sync-bxpjp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.543609 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxzxp\" (UniqueName: \"kubernetes.io/projected/c8ae1721-4ded-4e7d-9091-88ac908d0554-kube-api-access-pxzxp\") pod \"cinder-db-sync-bxpjp\" (UID: \"c8ae1721-4ded-4e7d-9091-88ac908d0554\") " pod="openstack/cinder-db-sync-bxpjp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.543926 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-mvljp"] Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.545029 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4163bfe8-0de7-4490-b825-55d0f0343787-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4163bfe8-0de7-4490-b825-55d0f0343787\") " pod="openstack/ceilometer-0" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.545148 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8ae1721-4ded-4e7d-9091-88ac908d0554-config-data\") pod \"cinder-db-sync-bxpjp\" (UID: \"c8ae1721-4ded-4e7d-9091-88ac908d0554\") " pod="openstack/cinder-db-sync-bxpjp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.545752 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4163bfe8-0de7-4490-b825-55d0f0343787-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4163bfe8-0de7-4490-b825-55d0f0343787\") " pod="openstack/ceilometer-0" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.546962 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4163bfe8-0de7-4490-b825-55d0f0343787-scripts\") pod \"ceilometer-0\" (UID: \"4163bfe8-0de7-4490-b825-55d0f0343787\") " pod="openstack/ceilometer-0" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.547791 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gf5bp\" (UniqueName: \"kubernetes.io/projected/4163bfe8-0de7-4490-b825-55d0f0343787-kube-api-access-gf5bp\") pod \"ceilometer-0\" (UID: \"4163bfe8-0de7-4490-b825-55d0f0343787\") " pod="openstack/ceilometer-0" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.551296 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d78ff46f5-kdzjr"] Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.560916 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4163bfe8-0de7-4490-b825-55d0f0343787-config-data\") pod \"ceilometer-0\" (UID: \"4163bfe8-0de7-4490-b825-55d0f0343787\") " pod="openstack/ceilometer-0" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.561355 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8ae1721-4ded-4e7d-9091-88ac908d0554-combined-ca-bundle\") pod \"cinder-db-sync-bxpjp\" (UID: \"c8ae1721-4ded-4e7d-9091-88ac908d0554\") " pod="openstack/cinder-db-sync-bxpjp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.578019 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.627161 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-ovsdbserver-sb\") pod \"dnsmasq-dns-d78ff46f5-kdzjr\" (UID: \"58898cc0-18e6-41c6-9ce2-149d081ff29b\") " pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.627226 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43aed302-5f95-4e51-b845-5a9cee9d1a08-combined-ca-bundle\") pod \"placement-db-sync-mvljp\" (UID: \"43aed302-5f95-4e51-b845-5a9cee9d1a08\") " pod="openstack/placement-db-sync-mvljp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.627261 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ad2af1c5-49df-4093-a38f-b1b47e2ddc1f-config\") pod \"neutron-db-sync-qgn9k\" (UID: \"ad2af1c5-49df-4093-a38f-b1b47e2ddc1f\") " pod="openstack/neutron-db-sync-qgn9k" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.627323 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2xnv\" (UniqueName: \"kubernetes.io/projected/ad2af1c5-49df-4093-a38f-b1b47e2ddc1f-kube-api-access-c2xnv\") pod \"neutron-db-sync-qgn9k\" (UID: \"ad2af1c5-49df-4093-a38f-b1b47e2ddc1f\") " pod="openstack/neutron-db-sync-qgn9k" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.627377 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-dns-svc\") pod \"dnsmasq-dns-d78ff46f5-kdzjr\" (UID: \"58898cc0-18e6-41c6-9ce2-149d081ff29b\") " pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.627398 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43aed302-5f95-4e51-b845-5a9cee9d1a08-scripts\") pod \"placement-db-sync-mvljp\" (UID: \"43aed302-5f95-4e51-b845-5a9cee9d1a08\") " pod="openstack/placement-db-sync-mvljp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.627442 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad2af1c5-49df-4093-a38f-b1b47e2ddc1f-combined-ca-bundle\") pod \"neutron-db-sync-qgn9k\" (UID: \"ad2af1c5-49df-4093-a38f-b1b47e2ddc1f\") " pod="openstack/neutron-db-sync-qgn9k" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.627461 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jn9c7\" (UniqueName: \"kubernetes.io/projected/b14d2018-932c-47b6-bba9-232dff3ce320-kube-api-access-jn9c7\") pod \"barbican-db-sync-k6zjg\" (UID: \"b14d2018-932c-47b6-bba9-232dff3ce320\") " pod="openstack/barbican-db-sync-k6zjg" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.627479 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-dns-swift-storage-0\") pod \"dnsmasq-dns-d78ff46f5-kdzjr\" (UID: \"58898cc0-18e6-41c6-9ce2-149d081ff29b\") " pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.627493 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m77lh\" (UniqueName: \"kubernetes.io/projected/58898cc0-18e6-41c6-9ce2-149d081ff29b-kube-api-access-m77lh\") pod \"dnsmasq-dns-d78ff46f5-kdzjr\" (UID: \"58898cc0-18e6-41c6-9ce2-149d081ff29b\") " pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.627536 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43aed302-5f95-4e51-b845-5a9cee9d1a08-config-data\") pod \"placement-db-sync-mvljp\" (UID: \"43aed302-5f95-4e51-b845-5a9cee9d1a08\") " pod="openstack/placement-db-sync-mvljp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.627575 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k557r\" (UniqueName: \"kubernetes.io/projected/43aed302-5f95-4e51-b845-5a9cee9d1a08-kube-api-access-k557r\") pod \"placement-db-sync-mvljp\" (UID: \"43aed302-5f95-4e51-b845-5a9cee9d1a08\") " pod="openstack/placement-db-sync-mvljp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.627603 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-config\") pod \"dnsmasq-dns-d78ff46f5-kdzjr\" (UID: \"58898cc0-18e6-41c6-9ce2-149d081ff29b\") " pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.627628 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43aed302-5f95-4e51-b845-5a9cee9d1a08-logs\") pod \"placement-db-sync-mvljp\" (UID: \"43aed302-5f95-4e51-b845-5a9cee9d1a08\") " pod="openstack/placement-db-sync-mvljp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.627709 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b14d2018-932c-47b6-bba9-232dff3ce320-db-sync-config-data\") pod \"barbican-db-sync-k6zjg\" (UID: \"b14d2018-932c-47b6-bba9-232dff3ce320\") " pod="openstack/barbican-db-sync-k6zjg" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.627751 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-ovsdbserver-nb\") pod \"dnsmasq-dns-d78ff46f5-kdzjr\" (UID: \"58898cc0-18e6-41c6-9ce2-149d081ff29b\") " pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.627961 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b14d2018-932c-47b6-bba9-232dff3ce320-combined-ca-bundle\") pod \"barbican-db-sync-k6zjg\" (UID: \"b14d2018-932c-47b6-bba9-232dff3ce320\") " pod="openstack/barbican-db-sync-k6zjg" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.631562 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-ovsdbserver-sb\") pod \"dnsmasq-dns-d78ff46f5-kdzjr\" (UID: \"58898cc0-18e6-41c6-9ce2-149d081ff29b\") " pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.633517 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-dns-svc\") pod \"dnsmasq-dns-d78ff46f5-kdzjr\" (UID: \"58898cc0-18e6-41c6-9ce2-149d081ff29b\") " pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.635046 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-dns-swift-storage-0\") pod \"dnsmasq-dns-d78ff46f5-kdzjr\" (UID: \"58898cc0-18e6-41c6-9ce2-149d081ff29b\") " pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.635313 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43aed302-5f95-4e51-b845-5a9cee9d1a08-logs\") pod \"placement-db-sync-mvljp\" (UID: \"43aed302-5f95-4e51-b845-5a9cee9d1a08\") " pod="openstack/placement-db-sync-mvljp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.635844 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-config\") pod \"dnsmasq-dns-d78ff46f5-kdzjr\" (UID: \"58898cc0-18e6-41c6-9ce2-149d081ff29b\") " pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.637986 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43aed302-5f95-4e51-b845-5a9cee9d1a08-combined-ca-bundle\") pod \"placement-db-sync-mvljp\" (UID: \"43aed302-5f95-4e51-b845-5a9cee9d1a08\") " pod="openstack/placement-db-sync-mvljp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.638921 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b14d2018-932c-47b6-bba9-232dff3ce320-db-sync-config-data\") pod \"barbican-db-sync-k6zjg\" (UID: \"b14d2018-932c-47b6-bba9-232dff3ce320\") " pod="openstack/barbican-db-sync-k6zjg" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.639359 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad2af1c5-49df-4093-a38f-b1b47e2ddc1f-combined-ca-bundle\") pod \"neutron-db-sync-qgn9k\" (UID: \"ad2af1c5-49df-4093-a38f-b1b47e2ddc1f\") " pod="openstack/neutron-db-sync-qgn9k" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.639442 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-ovsdbserver-nb\") pod \"dnsmasq-dns-d78ff46f5-kdzjr\" (UID: \"58898cc0-18e6-41c6-9ce2-149d081ff29b\") " pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.639812 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b14d2018-932c-47b6-bba9-232dff3ce320-combined-ca-bundle\") pod \"barbican-db-sync-k6zjg\" (UID: \"b14d2018-932c-47b6-bba9-232dff3ce320\") " pod="openstack/barbican-db-sync-k6zjg" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.641852 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ad2af1c5-49df-4093-a38f-b1b47e2ddc1f-config\") pod \"neutron-db-sync-qgn9k\" (UID: \"ad2af1c5-49df-4093-a38f-b1b47e2ddc1f\") " pod="openstack/neutron-db-sync-qgn9k" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.642797 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43aed302-5f95-4e51-b845-5a9cee9d1a08-scripts\") pod \"placement-db-sync-mvljp\" (UID: \"43aed302-5f95-4e51-b845-5a9cee9d1a08\") " pod="openstack/placement-db-sync-mvljp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.643831 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43aed302-5f95-4e51-b845-5a9cee9d1a08-config-data\") pod \"placement-db-sync-mvljp\" (UID: \"43aed302-5f95-4e51-b845-5a9cee9d1a08\") " pod="openstack/placement-db-sync-mvljp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.644740 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jn9c7\" (UniqueName: \"kubernetes.io/projected/b14d2018-932c-47b6-bba9-232dff3ce320-kube-api-access-jn9c7\") pod \"barbican-db-sync-k6zjg\" (UID: \"b14d2018-932c-47b6-bba9-232dff3ce320\") " pod="openstack/barbican-db-sync-k6zjg" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.647051 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2xnv\" (UniqueName: \"kubernetes.io/projected/ad2af1c5-49df-4093-a38f-b1b47e2ddc1f-kube-api-access-c2xnv\") pod \"neutron-db-sync-qgn9k\" (UID: \"ad2af1c5-49df-4093-a38f-b1b47e2ddc1f\") " pod="openstack/neutron-db-sync-qgn9k" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.649221 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m77lh\" (UniqueName: \"kubernetes.io/projected/58898cc0-18e6-41c6-9ce2-149d081ff29b-kube-api-access-m77lh\") pod \"dnsmasq-dns-d78ff46f5-kdzjr\" (UID: \"58898cc0-18e6-41c6-9ce2-149d081ff29b\") " pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.649603 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k557r\" (UniqueName: \"kubernetes.io/projected/43aed302-5f95-4e51-b845-5a9cee9d1a08-kube-api-access-k557r\") pod \"placement-db-sync-mvljp\" (UID: \"43aed302-5f95-4e51-b845-5a9cee9d1a08\") " pod="openstack/placement-db-sync-mvljp" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.655384 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" Nov 23 06:59:25 crc kubenswrapper[4559]: I1123 06:59:25.680061 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-bxpjp" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:25.770212 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nm8sn" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:25.899820 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-k6zjg" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:25.903097 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b54fd9f79-pp688"] Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:25.910021 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-qgn9k" Nov 23 06:59:26 crc kubenswrapper[4559]: W1123 06:59:25.914797 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa3c4b78_d545_4bf5_95cd_33ef6670e895.slice/crio-9ea74f55b239efee678cf0a51c56a1434239f6f0c5eb37ec3ec6482d63cc8561 WatchSource:0}: Error finding container 9ea74f55b239efee678cf0a51c56a1434239f6f0c5eb37ec3ec6482d63cc8561: Status 404 returned error can't find the container with id 9ea74f55b239efee678cf0a51c56a1434239f6f0c5eb37ec3ec6482d63cc8561 Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:25.927238 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-mvljp" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:25.993545 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.166835 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.167008 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.203825 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.205358 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.207195 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.208433 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.208448 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-tj89j" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.208605 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.218409 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.249843 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f34d682-a110-4d78-900b-da0ac3f9c5df-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.249913 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f34d682-a110-4d78-900b-da0ac3f9c5df-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.249947 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f34d682-a110-4d78-900b-da0ac3f9c5df-config-data\") pod \"glance-default-external-api-0\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.249983 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3f34d682-a110-4d78-900b-da0ac3f9c5df-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.249997 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f34d682-a110-4d78-900b-da0ac3f9c5df-scripts\") pod \"glance-default-external-api-0\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.250087 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzfns\" (UniqueName: \"kubernetes.io/projected/3f34d682-a110-4d78-900b-da0ac3f9c5df-kube-api-access-qzfns\") pod \"glance-default-external-api-0\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.250281 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f34d682-a110-4d78-900b-da0ac3f9c5df-logs\") pod \"glance-default-external-api-0\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.250365 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.262479 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.263832 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.265490 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.271243 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.282392 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.352410 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f34d682-a110-4d78-900b-da0ac3f9c5df-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.352448 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/089e3832-1bfa-4c6a-9747-994ec8c5ee36-logs\") pod \"glance-default-internal-api-0\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.352467 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f34d682-a110-4d78-900b-da0ac3f9c5df-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.352490 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f34d682-a110-4d78-900b-da0ac3f9c5df-config-data\") pod \"glance-default-external-api-0\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.352510 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3f34d682-a110-4d78-900b-da0ac3f9c5df-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.352526 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/089e3832-1bfa-4c6a-9747-994ec8c5ee36-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.352541 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f34d682-a110-4d78-900b-da0ac3f9c5df-scripts\") pod \"glance-default-external-api-0\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.352590 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzfns\" (UniqueName: \"kubernetes.io/projected/3f34d682-a110-4d78-900b-da0ac3f9c5df-kube-api-access-qzfns\") pod \"glance-default-external-api-0\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.352605 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/089e3832-1bfa-4c6a-9747-994ec8c5ee36-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.352653 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/089e3832-1bfa-4c6a-9747-994ec8c5ee36-config-data\") pod \"glance-default-internal-api-0\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.352702 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/089e3832-1bfa-4c6a-9747-994ec8c5ee36-scripts\") pod \"glance-default-internal-api-0\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.352760 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.352776 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f34d682-a110-4d78-900b-da0ac3f9c5df-logs\") pod \"glance-default-external-api-0\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.352829 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pbhm\" (UniqueName: \"kubernetes.io/projected/089e3832-1bfa-4c6a-9747-994ec8c5ee36-kube-api-access-7pbhm\") pod \"glance-default-internal-api-0\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.352846 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.352874 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/089e3832-1bfa-4c6a-9747-994ec8c5ee36-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.353280 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3f34d682-a110-4d78-900b-da0ac3f9c5df-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.353551 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f34d682-a110-4d78-900b-da0ac3f9c5df-logs\") pod \"glance-default-external-api-0\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.353962 4559 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-external-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.357079 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f34d682-a110-4d78-900b-da0ac3f9c5df-scripts\") pod \"glance-default-external-api-0\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.357349 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f34d682-a110-4d78-900b-da0ac3f9c5df-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.358657 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f34d682-a110-4d78-900b-da0ac3f9c5df-config-data\") pod \"glance-default-external-api-0\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.360513 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f34d682-a110-4d78-900b-da0ac3f9c5df-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.371280 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzfns\" (UniqueName: \"kubernetes.io/projected/3f34d682-a110-4d78-900b-da0ac3f9c5df-kube-api-access-qzfns\") pod \"glance-default-external-api-0\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.379325 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.454333 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/089e3832-1bfa-4c6a-9747-994ec8c5ee36-logs\") pod \"glance-default-internal-api-0\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.454393 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/089e3832-1bfa-4c6a-9747-994ec8c5ee36-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.454449 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/089e3832-1bfa-4c6a-9747-994ec8c5ee36-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.454488 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/089e3832-1bfa-4c6a-9747-994ec8c5ee36-config-data\") pod \"glance-default-internal-api-0\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.454559 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/089e3832-1bfa-4c6a-9747-994ec8c5ee36-scripts\") pod \"glance-default-internal-api-0\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.454625 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.454706 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pbhm\" (UniqueName: \"kubernetes.io/projected/089e3832-1bfa-4c6a-9747-994ec8c5ee36-kube-api-access-7pbhm\") pod \"glance-default-internal-api-0\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.454747 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/089e3832-1bfa-4c6a-9747-994ec8c5ee36-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.455036 4559 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.455238 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/089e3832-1bfa-4c6a-9747-994ec8c5ee36-logs\") pod \"glance-default-internal-api-0\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.455266 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/089e3832-1bfa-4c6a-9747-994ec8c5ee36-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.457585 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/089e3832-1bfa-4c6a-9747-994ec8c5ee36-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.458940 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/089e3832-1bfa-4c6a-9747-994ec8c5ee36-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.459263 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/089e3832-1bfa-4c6a-9747-994ec8c5ee36-scripts\") pod \"glance-default-internal-api-0\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.464800 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/089e3832-1bfa-4c6a-9747-994ec8c5ee36-config-data\") pod \"glance-default-internal-api-0\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.470156 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pbhm\" (UniqueName: \"kubernetes.io/projected/089e3832-1bfa-4c6a-9747-994ec8c5ee36-kube-api-access-7pbhm\") pod \"glance-default-internal-api-0\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.483953 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.521350 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.587423 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-mvljp"] Nov 23 06:59:26 crc kubenswrapper[4559]: W1123 06:59:26.589186 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43aed302_5f95_4e51_b845_5a9cee9d1a08.slice/crio-357376614a62e7b787fddd897553a74c3cb166227d16413f80afe94f71321ae9 WatchSource:0}: Error finding container 357376614a62e7b787fddd897553a74c3cb166227d16413f80afe94f71321ae9: Status 404 returned error can't find the container with id 357376614a62e7b787fddd897553a74c3cb166227d16413f80afe94f71321ae9 Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.589296 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.813805 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-qgn9k"] Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.839935 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nm8sn"] Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.855081 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-bxpjp"] Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.871670 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d78ff46f5-kdzjr"] Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.877238 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-k6zjg"] Nov 23 06:59:26 crc kubenswrapper[4559]: W1123 06:59:26.877859 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58898cc0_18e6_41c6_9ce2_149d081ff29b.slice/crio-c9e7ab3cf0b100c155267ee1c988ad1cc79bc7412ea16b6da38e6ceb7c86a95f WatchSource:0}: Error finding container c9e7ab3cf0b100c155267ee1c988ad1cc79bc7412ea16b6da38e6ceb7c86a95f: Status 404 returned error can't find the container with id c9e7ab3cf0b100c155267ee1c988ad1cc79bc7412ea16b6da38e6ceb7c86a95f Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.905397 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4163bfe8-0de7-4490-b825-55d0f0343787","Type":"ContainerStarted","Data":"abd76a57cf9f3e9befff8730283bb9ea4c4f010c2d5207130c0f760b6cd1fa84"} Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.909423 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-bxpjp" event={"ID":"c8ae1721-4ded-4e7d-9091-88ac908d0554","Type":"ContainerStarted","Data":"6a9113470a7a2a0237d3f88cfdd768d72e140dfacc567addaaea7488009a56d6"} Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.914372 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-k6zjg" event={"ID":"b14d2018-932c-47b6-bba9-232dff3ce320","Type":"ContainerStarted","Data":"57f3f54f0ed9b335c62666e41423bd0c550cd93e89fb3684a1109a89c6cfd699"} Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.915578 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-mvljp" event={"ID":"43aed302-5f95-4e51-b845-5a9cee9d1a08","Type":"ContainerStarted","Data":"357376614a62e7b787fddd897553a74c3cb166227d16413f80afe94f71321ae9"} Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.917305 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-qgn9k" event={"ID":"ad2af1c5-49df-4093-a38f-b1b47e2ddc1f","Type":"ContainerStarted","Data":"b4c12b5edf1d8c7628b432bcaa7f8f25b3a5fe533e3494c119d22e35af55d52c"} Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.925836 4559 generic.go:334] "Generic (PLEG): container finished" podID="aa3c4b78-d545-4bf5-95cd-33ef6670e895" containerID="dda007d0aec9da7c9b50cbdf63e1e4f5ebbd715b09d521237ea01cc7d61d6eec" exitCode=0 Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.925977 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b54fd9f79-pp688" event={"ID":"aa3c4b78-d545-4bf5-95cd-33ef6670e895","Type":"ContainerDied","Data":"dda007d0aec9da7c9b50cbdf63e1e4f5ebbd715b09d521237ea01cc7d61d6eec"} Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.926021 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b54fd9f79-pp688" event={"ID":"aa3c4b78-d545-4bf5-95cd-33ef6670e895","Type":"ContainerStarted","Data":"9ea74f55b239efee678cf0a51c56a1434239f6f0c5eb37ec3ec6482d63cc8561"} Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.927369 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nm8sn" event={"ID":"18ba7210-7560-48f6-b164-780a6c84ef5f","Type":"ContainerStarted","Data":"a2e18a3af5dea13662fe732bcb7f59b4dc12562ffc1a0148b5cb55162c6bebc3"} Nov 23 06:59:26 crc kubenswrapper[4559]: I1123 06:59:26.931885 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" event={"ID":"58898cc0-18e6-41c6-9ce2-149d081ff29b","Type":"ContainerStarted","Data":"c9e7ab3cf0b100c155267ee1c988ad1cc79bc7412ea16b6da38e6ceb7c86a95f"} Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.152185 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.183700 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.233714 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b54fd9f79-pp688" Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.270322 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.270495 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-dns-swift-storage-0\") pod \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\" (UID: \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\") " Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.270537 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-ovsdbserver-nb\") pod \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\" (UID: \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\") " Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.270654 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-config\") pod \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\" (UID: \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\") " Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.270697 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qn4rt\" (UniqueName: \"kubernetes.io/projected/aa3c4b78-d545-4bf5-95cd-33ef6670e895-kube-api-access-qn4rt\") pod \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\" (UID: \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\") " Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.270737 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-ovsdbserver-sb\") pod \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\" (UID: \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\") " Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.270774 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-dns-svc\") pod \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\" (UID: \"aa3c4b78-d545-4bf5-95cd-33ef6670e895\") " Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.325454 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "aa3c4b78-d545-4bf5-95cd-33ef6670e895" (UID: "aa3c4b78-d545-4bf5-95cd-33ef6670e895"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.338079 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa3c4b78-d545-4bf5-95cd-33ef6670e895-kube-api-access-qn4rt" (OuterVolumeSpecName: "kube-api-access-qn4rt") pod "aa3c4b78-d545-4bf5-95cd-33ef6670e895" (UID: "aa3c4b78-d545-4bf5-95cd-33ef6670e895"). InnerVolumeSpecName "kube-api-access-qn4rt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.346267 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "aa3c4b78-d545-4bf5-95cd-33ef6670e895" (UID: "aa3c4b78-d545-4bf5-95cd-33ef6670e895"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.347076 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-config" (OuterVolumeSpecName: "config") pod "aa3c4b78-d545-4bf5-95cd-33ef6670e895" (UID: "aa3c4b78-d545-4bf5-95cd-33ef6670e895"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.380403 4559 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.380431 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.380441 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qn4rt\" (UniqueName: \"kubernetes.io/projected/aa3c4b78-d545-4bf5-95cd-33ef6670e895-kube-api-access-qn4rt\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.380453 4559 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.386164 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aa3c4b78-d545-4bf5-95cd-33ef6670e895" (UID: "aa3c4b78-d545-4bf5-95cd-33ef6670e895"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.400041 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.405758 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.464447 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aa3c4b78-d545-4bf5-95cd-33ef6670e895" (UID: "aa3c4b78-d545-4bf5-95cd-33ef6670e895"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.483016 4559 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.483047 4559 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa3c4b78-d545-4bf5-95cd-33ef6670e895-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.946250 4559 generic.go:334] "Generic (PLEG): container finished" podID="58898cc0-18e6-41c6-9ce2-149d081ff29b" containerID="75879545707b8bec91ac310eb56a716157422ee90fb701af77ed451048b5e4c9" exitCode=0 Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.946351 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" event={"ID":"58898cc0-18e6-41c6-9ce2-149d081ff29b","Type":"ContainerDied","Data":"75879545707b8bec91ac310eb56a716157422ee90fb701af77ed451048b5e4c9"} Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.949783 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3f34d682-a110-4d78-900b-da0ac3f9c5df","Type":"ContainerStarted","Data":"c2ea7bf31dca742699dbe147d3090a60a159589e90db40c0601b82553039293b"} Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.949821 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3f34d682-a110-4d78-900b-da0ac3f9c5df","Type":"ContainerStarted","Data":"8848e494fd0ef9ea57c726f4ecf38128819c5796252d91090f2135d24928795d"} Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.955832 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-qgn9k" event={"ID":"ad2af1c5-49df-4093-a38f-b1b47e2ddc1f","Type":"ContainerStarted","Data":"bbd5eebeecbdf13f45dd156157392985c3a4a276178b65794a062e0829d3d543"} Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.958029 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b54fd9f79-pp688" event={"ID":"aa3c4b78-d545-4bf5-95cd-33ef6670e895","Type":"ContainerDied","Data":"9ea74f55b239efee678cf0a51c56a1434239f6f0c5eb37ec3ec6482d63cc8561"} Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.958065 4559 scope.go:117] "RemoveContainer" containerID="dda007d0aec9da7c9b50cbdf63e1e4f5ebbd715b09d521237ea01cc7d61d6eec" Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.958149 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b54fd9f79-pp688" Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.983285 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"089e3832-1bfa-4c6a-9747-994ec8c5ee36","Type":"ContainerStarted","Data":"de73f9f96f9793870a649d6ea8aaacd847dd180df529f2b4d9378582b2364b33"} Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.986599 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-qgn9k" podStartSLOduration=2.986582779 podStartE2EDuration="2.986582779s" podCreationTimestamp="2025-11-23 06:59:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:59:27.983482569 +0000 UTC m=+870.005468183" watchObservedRunningTime="2025-11-23 06:59:27.986582779 +0000 UTC m=+870.008568393" Nov 23 06:59:27 crc kubenswrapper[4559]: I1123 06:59:27.995362 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nm8sn" event={"ID":"18ba7210-7560-48f6-b164-780a6c84ef5f","Type":"ContainerStarted","Data":"8f7af979fa7061247fe8031e3f0c0c907f89b6bdedef6766b14c7cde78cb28fd"} Nov 23 06:59:28 crc kubenswrapper[4559]: I1123 06:59:28.035602 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b54fd9f79-pp688"] Nov 23 06:59:28 crc kubenswrapper[4559]: I1123 06:59:28.041014 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b54fd9f79-pp688"] Nov 23 06:59:28 crc kubenswrapper[4559]: I1123 06:59:28.041389 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-nm8sn" podStartSLOduration=3.041373245 podStartE2EDuration="3.041373245s" podCreationTimestamp="2025-11-23 06:59:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:59:28.023871269 +0000 UTC m=+870.045856882" watchObservedRunningTime="2025-11-23 06:59:28.041373245 +0000 UTC m=+870.063358859" Nov 23 06:59:28 crc kubenswrapper[4559]: I1123 06:59:28.302118 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa3c4b78-d545-4bf5-95cd-33ef6670e895" path="/var/lib/kubelet/pods/aa3c4b78-d545-4bf5-95cd-33ef6670e895/volumes" Nov 23 06:59:29 crc kubenswrapper[4559]: I1123 06:59:29.036469 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" event={"ID":"58898cc0-18e6-41c6-9ce2-149d081ff29b","Type":"ContainerStarted","Data":"93860435ee52af0972f94019211071574356bbc2034f1651ff6d61882b1e7bbe"} Nov 23 06:59:29 crc kubenswrapper[4559]: I1123 06:59:29.037938 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" Nov 23 06:59:29 crc kubenswrapper[4559]: I1123 06:59:29.043020 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"089e3832-1bfa-4c6a-9747-994ec8c5ee36","Type":"ContainerStarted","Data":"8959d6d6aef2525e4f4a2b37081ac1dd1152197e5713d7c5744ce02fcd4db956"} Nov 23 06:59:29 crc kubenswrapper[4559]: I1123 06:59:29.056307 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" podStartSLOduration=4.056290832 podStartE2EDuration="4.056290832s" podCreationTimestamp="2025-11-23 06:59:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:59:29.052738782 +0000 UTC m=+871.074724396" watchObservedRunningTime="2025-11-23 06:59:29.056290832 +0000 UTC m=+871.078276446" Nov 23 06:59:30 crc kubenswrapper[4559]: I1123 06:59:30.058811 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"089e3832-1bfa-4c6a-9747-994ec8c5ee36","Type":"ContainerStarted","Data":"a9c0b0f50c2fe99c763c859074d3ffad9c637689755bd5ed8134ae92fe36451e"} Nov 23 06:59:30 crc kubenswrapper[4559]: I1123 06:59:30.059002 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="089e3832-1bfa-4c6a-9747-994ec8c5ee36" containerName="glance-log" containerID="cri-o://8959d6d6aef2525e4f4a2b37081ac1dd1152197e5713d7c5744ce02fcd4db956" gracePeriod=30 Nov 23 06:59:30 crc kubenswrapper[4559]: I1123 06:59:30.059403 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="089e3832-1bfa-4c6a-9747-994ec8c5ee36" containerName="glance-httpd" containerID="cri-o://a9c0b0f50c2fe99c763c859074d3ffad9c637689755bd5ed8134ae92fe36451e" gracePeriod=30 Nov 23 06:59:30 crc kubenswrapper[4559]: I1123 06:59:30.066899 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3f34d682-a110-4d78-900b-da0ac3f9c5df","Type":"ContainerStarted","Data":"93039f377356277538dfdf8cb1477bb70bbfdd2cccd38bb94b3f0f13bca76fcd"} Nov 23 06:59:30 crc kubenswrapper[4559]: I1123 06:59:30.066954 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3f34d682-a110-4d78-900b-da0ac3f9c5df" containerName="glance-log" containerID="cri-o://c2ea7bf31dca742699dbe147d3090a60a159589e90db40c0601b82553039293b" gracePeriod=30 Nov 23 06:59:30 crc kubenswrapper[4559]: I1123 06:59:30.067031 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3f34d682-a110-4d78-900b-da0ac3f9c5df" containerName="glance-httpd" containerID="cri-o://93039f377356277538dfdf8cb1477bb70bbfdd2cccd38bb94b3f0f13bca76fcd" gracePeriod=30 Nov 23 06:59:30 crc kubenswrapper[4559]: I1123 06:59:30.081936 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.081924554 podStartE2EDuration="5.081924554s" podCreationTimestamp="2025-11-23 06:59:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:59:30.08090133 +0000 UTC m=+872.102886943" watchObservedRunningTime="2025-11-23 06:59:30.081924554 +0000 UTC m=+872.103910178" Nov 23 06:59:30 crc kubenswrapper[4559]: I1123 06:59:30.104732 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.104719144 podStartE2EDuration="5.104719144s" podCreationTimestamp="2025-11-23 06:59:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:59:30.09812893 +0000 UTC m=+872.120114544" watchObservedRunningTime="2025-11-23 06:59:30.104719144 +0000 UTC m=+872.126704758" Nov 23 06:59:31 crc kubenswrapper[4559]: I1123 06:59:31.078289 4559 generic.go:334] "Generic (PLEG): container finished" podID="3f34d682-a110-4d78-900b-da0ac3f9c5df" containerID="93039f377356277538dfdf8cb1477bb70bbfdd2cccd38bb94b3f0f13bca76fcd" exitCode=143 Nov 23 06:59:31 crc kubenswrapper[4559]: I1123 06:59:31.078552 4559 generic.go:334] "Generic (PLEG): container finished" podID="3f34d682-a110-4d78-900b-da0ac3f9c5df" containerID="c2ea7bf31dca742699dbe147d3090a60a159589e90db40c0601b82553039293b" exitCode=143 Nov 23 06:59:31 crc kubenswrapper[4559]: I1123 06:59:31.078356 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3f34d682-a110-4d78-900b-da0ac3f9c5df","Type":"ContainerDied","Data":"93039f377356277538dfdf8cb1477bb70bbfdd2cccd38bb94b3f0f13bca76fcd"} Nov 23 06:59:31 crc kubenswrapper[4559]: I1123 06:59:31.078625 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3f34d682-a110-4d78-900b-da0ac3f9c5df","Type":"ContainerDied","Data":"c2ea7bf31dca742699dbe147d3090a60a159589e90db40c0601b82553039293b"} Nov 23 06:59:31 crc kubenswrapper[4559]: I1123 06:59:31.082275 4559 generic.go:334] "Generic (PLEG): container finished" podID="089e3832-1bfa-4c6a-9747-994ec8c5ee36" containerID="a9c0b0f50c2fe99c763c859074d3ffad9c637689755bd5ed8134ae92fe36451e" exitCode=143 Nov 23 06:59:31 crc kubenswrapper[4559]: I1123 06:59:31.082303 4559 generic.go:334] "Generic (PLEG): container finished" podID="089e3832-1bfa-4c6a-9747-994ec8c5ee36" containerID="8959d6d6aef2525e4f4a2b37081ac1dd1152197e5713d7c5744ce02fcd4db956" exitCode=143 Nov 23 06:59:31 crc kubenswrapper[4559]: I1123 06:59:31.082319 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"089e3832-1bfa-4c6a-9747-994ec8c5ee36","Type":"ContainerDied","Data":"a9c0b0f50c2fe99c763c859074d3ffad9c637689755bd5ed8134ae92fe36451e"} Nov 23 06:59:31 crc kubenswrapper[4559]: I1123 06:59:31.082366 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"089e3832-1bfa-4c6a-9747-994ec8c5ee36","Type":"ContainerDied","Data":"8959d6d6aef2525e4f4a2b37081ac1dd1152197e5713d7c5744ce02fcd4db956"} Nov 23 06:59:31 crc kubenswrapper[4559]: I1123 06:59:31.084210 4559 generic.go:334] "Generic (PLEG): container finished" podID="18ba7210-7560-48f6-b164-780a6c84ef5f" containerID="8f7af979fa7061247fe8031e3f0c0c907f89b6bdedef6766b14c7cde78cb28fd" exitCode=0 Nov 23 06:59:31 crc kubenswrapper[4559]: I1123 06:59:31.084291 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nm8sn" event={"ID":"18ba7210-7560-48f6-b164-780a6c84ef5f","Type":"ContainerDied","Data":"8f7af979fa7061247fe8031e3f0c0c907f89b6bdedef6766b14c7cde78cb28fd"} Nov 23 06:59:34 crc kubenswrapper[4559]: I1123 06:59:34.108068 4559 generic.go:334] "Generic (PLEG): container finished" podID="ad2af1c5-49df-4093-a38f-b1b47e2ddc1f" containerID="bbd5eebeecbdf13f45dd156157392985c3a4a276178b65794a062e0829d3d543" exitCode=0 Nov 23 06:59:34 crc kubenswrapper[4559]: I1123 06:59:34.108142 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-qgn9k" event={"ID":"ad2af1c5-49df-4093-a38f-b1b47e2ddc1f","Type":"ContainerDied","Data":"bbd5eebeecbdf13f45dd156157392985c3a4a276178b65794a062e0829d3d543"} Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.464776 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.468611 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-qgn9k" Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.478309 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3f34d682-a110-4d78-900b-da0ac3f9c5df-httpd-run\") pod \"3f34d682-a110-4d78-900b-da0ac3f9c5df\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.478381 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad2af1c5-49df-4093-a38f-b1b47e2ddc1f-combined-ca-bundle\") pod \"ad2af1c5-49df-4093-a38f-b1b47e2ddc1f\" (UID: \"ad2af1c5-49df-4093-a38f-b1b47e2ddc1f\") " Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.478433 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"3f34d682-a110-4d78-900b-da0ac3f9c5df\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.478503 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2xnv\" (UniqueName: \"kubernetes.io/projected/ad2af1c5-49df-4093-a38f-b1b47e2ddc1f-kube-api-access-c2xnv\") pod \"ad2af1c5-49df-4093-a38f-b1b47e2ddc1f\" (UID: \"ad2af1c5-49df-4093-a38f-b1b47e2ddc1f\") " Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.478561 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f34d682-a110-4d78-900b-da0ac3f9c5df-logs\") pod \"3f34d682-a110-4d78-900b-da0ac3f9c5df\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.478622 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f34d682-a110-4d78-900b-da0ac3f9c5df-config-data\") pod \"3f34d682-a110-4d78-900b-da0ac3f9c5df\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.478977 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f34d682-a110-4d78-900b-da0ac3f9c5df-logs" (OuterVolumeSpecName: "logs") pod "3f34d682-a110-4d78-900b-da0ac3f9c5df" (UID: "3f34d682-a110-4d78-900b-da0ac3f9c5df"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.479342 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ad2af1c5-49df-4093-a38f-b1b47e2ddc1f-config\") pod \"ad2af1c5-49df-4093-a38f-b1b47e2ddc1f\" (UID: \"ad2af1c5-49df-4093-a38f-b1b47e2ddc1f\") " Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.479403 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f34d682-a110-4d78-900b-da0ac3f9c5df-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3f34d682-a110-4d78-900b-da0ac3f9c5df" (UID: "3f34d682-a110-4d78-900b-da0ac3f9c5df"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.481506 4559 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3f34d682-a110-4d78-900b-da0ac3f9c5df-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.481831 4559 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f34d682-a110-4d78-900b-da0ac3f9c5df-logs\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.493529 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad2af1c5-49df-4093-a38f-b1b47e2ddc1f-kube-api-access-c2xnv" (OuterVolumeSpecName: "kube-api-access-c2xnv") pod "ad2af1c5-49df-4093-a38f-b1b47e2ddc1f" (UID: "ad2af1c5-49df-4093-a38f-b1b47e2ddc1f"). InnerVolumeSpecName "kube-api-access-c2xnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.496962 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "3f34d682-a110-4d78-900b-da0ac3f9c5df" (UID: "3f34d682-a110-4d78-900b-da0ac3f9c5df"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.517924 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad2af1c5-49df-4093-a38f-b1b47e2ddc1f-config" (OuterVolumeSpecName: "config") pod "ad2af1c5-49df-4093-a38f-b1b47e2ddc1f" (UID: "ad2af1c5-49df-4093-a38f-b1b47e2ddc1f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.531007 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad2af1c5-49df-4093-a38f-b1b47e2ddc1f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ad2af1c5-49df-4093-a38f-b1b47e2ddc1f" (UID: "ad2af1c5-49df-4093-a38f-b1b47e2ddc1f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.553523 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f34d682-a110-4d78-900b-da0ac3f9c5df-config-data" (OuterVolumeSpecName: "config-data") pod "3f34d682-a110-4d78-900b-da0ac3f9c5df" (UID: "3f34d682-a110-4d78-900b-da0ac3f9c5df"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.583699 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f34d682-a110-4d78-900b-da0ac3f9c5df-combined-ca-bundle\") pod \"3f34d682-a110-4d78-900b-da0ac3f9c5df\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.583746 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f34d682-a110-4d78-900b-da0ac3f9c5df-public-tls-certs\") pod \"3f34d682-a110-4d78-900b-da0ac3f9c5df\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.583823 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzfns\" (UniqueName: \"kubernetes.io/projected/3f34d682-a110-4d78-900b-da0ac3f9c5df-kube-api-access-qzfns\") pod \"3f34d682-a110-4d78-900b-da0ac3f9c5df\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.584065 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f34d682-a110-4d78-900b-da0ac3f9c5df-scripts\") pod \"3f34d682-a110-4d78-900b-da0ac3f9c5df\" (UID: \"3f34d682-a110-4d78-900b-da0ac3f9c5df\") " Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.584605 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad2af1c5-49df-4093-a38f-b1b47e2ddc1f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.584660 4559 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.584673 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2xnv\" (UniqueName: \"kubernetes.io/projected/ad2af1c5-49df-4093-a38f-b1b47e2ddc1f-kube-api-access-c2xnv\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.584682 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f34d682-a110-4d78-900b-da0ac3f9c5df-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.584691 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/ad2af1c5-49df-4093-a38f-b1b47e2ddc1f-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.591113 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f34d682-a110-4d78-900b-da0ac3f9c5df-scripts" (OuterVolumeSpecName: "scripts") pod "3f34d682-a110-4d78-900b-da0ac3f9c5df" (UID: "3f34d682-a110-4d78-900b-da0ac3f9c5df"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.593294 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f34d682-a110-4d78-900b-da0ac3f9c5df-kube-api-access-qzfns" (OuterVolumeSpecName: "kube-api-access-qzfns") pod "3f34d682-a110-4d78-900b-da0ac3f9c5df" (UID: "3f34d682-a110-4d78-900b-da0ac3f9c5df"). InnerVolumeSpecName "kube-api-access-qzfns". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.601821 4559 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.604449 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f34d682-a110-4d78-900b-da0ac3f9c5df-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3f34d682-a110-4d78-900b-da0ac3f9c5df" (UID: "3f34d682-a110-4d78-900b-da0ac3f9c5df"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.635083 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f34d682-a110-4d78-900b-da0ac3f9c5df-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "3f34d682-a110-4d78-900b-da0ac3f9c5df" (UID: "3f34d682-a110-4d78-900b-da0ac3f9c5df"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.657969 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.686002 4559 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f34d682-a110-4d78-900b-da0ac3f9c5df-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.686033 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f34d682-a110-4d78-900b-da0ac3f9c5df-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.686043 4559 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f34d682-a110-4d78-900b-da0ac3f9c5df-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.686054 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzfns\" (UniqueName: \"kubernetes.io/projected/3f34d682-a110-4d78-900b-da0ac3f9c5df-kube-api-access-qzfns\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.686065 4559 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.729010 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-648ff47655-pcz6l"] Nov 23 06:59:35 crc kubenswrapper[4559]: I1123 06:59:35.729734 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-648ff47655-pcz6l" podUID="9dafbd81-21bc-4d7e-a61d-ba694502490c" containerName="dnsmasq-dns" containerID="cri-o://b87622aa82b32ff182c836aa8a7f812dd11a95f19917e11e867c5184307dc0c6" gracePeriod=10 Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.068462 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nm8sn" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.073250 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.168865 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-qgn9k" event={"ID":"ad2af1c5-49df-4093-a38f-b1b47e2ddc1f","Type":"ContainerDied","Data":"b4c12b5edf1d8c7628b432bcaa7f8f25b3a5fe533e3494c119d22e35af55d52c"} Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.169115 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4c12b5edf1d8c7628b432bcaa7f8f25b3a5fe533e3494c119d22e35af55d52c" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.169293 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-qgn9k" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.199746 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"089e3832-1bfa-4c6a-9747-994ec8c5ee36","Type":"ContainerDied","Data":"de73f9f96f9793870a649d6ea8aaacd847dd180df529f2b4d9378582b2364b33"} Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.199879 4559 scope.go:117] "RemoveContainer" containerID="a9c0b0f50c2fe99c763c859074d3ffad9c637689755bd5ed8134ae92fe36451e" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.200038 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.204086 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nm8sn" event={"ID":"18ba7210-7560-48f6-b164-780a6c84ef5f","Type":"ContainerDied","Data":"a2e18a3af5dea13662fe732bcb7f59b4dc12562ffc1a0148b5cb55162c6bebc3"} Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.204181 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2e18a3af5dea13662fe732bcb7f59b4dc12562ffc1a0148b5cb55162c6bebc3" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.204271 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nm8sn" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.210295 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzs77\" (UniqueName: \"kubernetes.io/projected/18ba7210-7560-48f6-b164-780a6c84ef5f-kube-api-access-vzs77\") pod \"18ba7210-7560-48f6-b164-780a6c84ef5f\" (UID: \"18ba7210-7560-48f6-b164-780a6c84ef5f\") " Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.210333 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-combined-ca-bundle\") pod \"18ba7210-7560-48f6-b164-780a6c84ef5f\" (UID: \"18ba7210-7560-48f6-b164-780a6c84ef5f\") " Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.210371 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/089e3832-1bfa-4c6a-9747-994ec8c5ee36-internal-tls-certs\") pod \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.210398 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pbhm\" (UniqueName: \"kubernetes.io/projected/089e3832-1bfa-4c6a-9747-994ec8c5ee36-kube-api-access-7pbhm\") pod \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.210443 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.210470 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/089e3832-1bfa-4c6a-9747-994ec8c5ee36-logs\") pod \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.210485 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/089e3832-1bfa-4c6a-9747-994ec8c5ee36-config-data\") pod \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.210504 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-config-data\") pod \"18ba7210-7560-48f6-b164-780a6c84ef5f\" (UID: \"18ba7210-7560-48f6-b164-780a6c84ef5f\") " Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.210546 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-credential-keys\") pod \"18ba7210-7560-48f6-b164-780a6c84ef5f\" (UID: \"18ba7210-7560-48f6-b164-780a6c84ef5f\") " Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.210564 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-scripts\") pod \"18ba7210-7560-48f6-b164-780a6c84ef5f\" (UID: \"18ba7210-7560-48f6-b164-780a6c84ef5f\") " Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.210600 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/089e3832-1bfa-4c6a-9747-994ec8c5ee36-combined-ca-bundle\") pod \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.210623 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/089e3832-1bfa-4c6a-9747-994ec8c5ee36-httpd-run\") pod \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.210652 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/089e3832-1bfa-4c6a-9747-994ec8c5ee36-scripts\") pod \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\" (UID: \"089e3832-1bfa-4c6a-9747-994ec8c5ee36\") " Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.210684 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-fernet-keys\") pod \"18ba7210-7560-48f6-b164-780a6c84ef5f\" (UID: \"18ba7210-7560-48f6-b164-780a6c84ef5f\") " Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.223288 4559 generic.go:334] "Generic (PLEG): container finished" podID="9dafbd81-21bc-4d7e-a61d-ba694502490c" containerID="b87622aa82b32ff182c836aa8a7f812dd11a95f19917e11e867c5184307dc0c6" exitCode=0 Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.223402 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-648ff47655-pcz6l" event={"ID":"9dafbd81-21bc-4d7e-a61d-ba694502490c","Type":"ContainerDied","Data":"b87622aa82b32ff182c836aa8a7f812dd11a95f19917e11e867c5184307dc0c6"} Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.226209 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/089e3832-1bfa-4c6a-9747-994ec8c5ee36-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "089e3832-1bfa-4c6a-9747-994ec8c5ee36" (UID: "089e3832-1bfa-4c6a-9747-994ec8c5ee36"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.230059 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/089e3832-1bfa-4c6a-9747-994ec8c5ee36-logs" (OuterVolumeSpecName: "logs") pod "089e3832-1bfa-4c6a-9747-994ec8c5ee36" (UID: "089e3832-1bfa-4c6a-9747-994ec8c5ee36"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.243006 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "18ba7210-7560-48f6-b164-780a6c84ef5f" (UID: "18ba7210-7560-48f6-b164-780a6c84ef5f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.267101 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/089e3832-1bfa-4c6a-9747-994ec8c5ee36-scripts" (OuterVolumeSpecName: "scripts") pod "089e3832-1bfa-4c6a-9747-994ec8c5ee36" (UID: "089e3832-1bfa-4c6a-9747-994ec8c5ee36"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.267174 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-scripts" (OuterVolumeSpecName: "scripts") pod "18ba7210-7560-48f6-b164-780a6c84ef5f" (UID: "18ba7210-7560-48f6-b164-780a6c84ef5f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.267254 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18ba7210-7560-48f6-b164-780a6c84ef5f-kube-api-access-vzs77" (OuterVolumeSpecName: "kube-api-access-vzs77") pod "18ba7210-7560-48f6-b164-780a6c84ef5f" (UID: "18ba7210-7560-48f6-b164-780a6c84ef5f"). InnerVolumeSpecName "kube-api-access-vzs77". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.267309 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "089e3832-1bfa-4c6a-9747-994ec8c5ee36" (UID: "089e3832-1bfa-4c6a-9747-994ec8c5ee36"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.286304 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/089e3832-1bfa-4c6a-9747-994ec8c5ee36-kube-api-access-7pbhm" (OuterVolumeSpecName: "kube-api-access-7pbhm") pod "089e3832-1bfa-4c6a-9747-994ec8c5ee36" (UID: "089e3832-1bfa-4c6a-9747-994ec8c5ee36"). InnerVolumeSpecName "kube-api-access-7pbhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.286368 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.287496 4559 scope.go:117] "RemoveContainer" containerID="8959d6d6aef2525e4f4a2b37081ac1dd1152197e5713d7c5744ce02fcd4db956" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.293558 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "18ba7210-7560-48f6-b164-780a6c84ef5f" (UID: "18ba7210-7560-48f6-b164-780a6c84ef5f"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.314950 4559 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.314977 4559 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/089e3832-1bfa-4c6a-9747-994ec8c5ee36-logs\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.314990 4559 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.314999 4559 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.315008 4559 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/089e3832-1bfa-4c6a-9747-994ec8c5ee36-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.315016 4559 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/089e3832-1bfa-4c6a-9747-994ec8c5ee36-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.315024 4559 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.315031 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzs77\" (UniqueName: \"kubernetes.io/projected/18ba7210-7560-48f6-b164-780a6c84ef5f-kube-api-access-vzs77\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.315040 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pbhm\" (UniqueName: \"kubernetes.io/projected/089e3832-1bfa-4c6a-9747-994ec8c5ee36-kube-api-access-7pbhm\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.316447 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3f34d682-a110-4d78-900b-da0ac3f9c5df","Type":"ContainerDied","Data":"8848e494fd0ef9ea57c726f4ecf38128819c5796252d91090f2135d24928795d"} Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.428127 4559 scope.go:117] "RemoveContainer" containerID="93039f377356277538dfdf8cb1477bb70bbfdd2cccd38bb94b3f0f13bca76fcd" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.428602 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.487933 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-config-data" (OuterVolumeSpecName: "config-data") pod "18ba7210-7560-48f6-b164-780a6c84ef5f" (UID: "18ba7210-7560-48f6-b164-780a6c84ef5f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.489857 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.492202 4559 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.518092 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 06:59:36 crc kubenswrapper[4559]: E1123 06:59:36.518546 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad2af1c5-49df-4093-a38f-b1b47e2ddc1f" containerName="neutron-db-sync" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.518563 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad2af1c5-49df-4093-a38f-b1b47e2ddc1f" containerName="neutron-db-sync" Nov 23 06:59:36 crc kubenswrapper[4559]: E1123 06:59:36.518582 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f34d682-a110-4d78-900b-da0ac3f9c5df" containerName="glance-log" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.518588 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f34d682-a110-4d78-900b-da0ac3f9c5df" containerName="glance-log" Nov 23 06:59:36 crc kubenswrapper[4559]: E1123 06:59:36.518603 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa3c4b78-d545-4bf5-95cd-33ef6670e895" containerName="init" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.518611 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa3c4b78-d545-4bf5-95cd-33ef6670e895" containerName="init" Nov 23 06:59:36 crc kubenswrapper[4559]: E1123 06:59:36.518620 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="089e3832-1bfa-4c6a-9747-994ec8c5ee36" containerName="glance-log" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.518626 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="089e3832-1bfa-4c6a-9747-994ec8c5ee36" containerName="glance-log" Nov 23 06:59:36 crc kubenswrapper[4559]: E1123 06:59:36.518652 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="089e3832-1bfa-4c6a-9747-994ec8c5ee36" containerName="glance-httpd" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.518658 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="089e3832-1bfa-4c6a-9747-994ec8c5ee36" containerName="glance-httpd" Nov 23 06:59:36 crc kubenswrapper[4559]: E1123 06:59:36.518667 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18ba7210-7560-48f6-b164-780a6c84ef5f" containerName="keystone-bootstrap" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.518674 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="18ba7210-7560-48f6-b164-780a6c84ef5f" containerName="keystone-bootstrap" Nov 23 06:59:36 crc kubenswrapper[4559]: E1123 06:59:36.518684 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f34d682-a110-4d78-900b-da0ac3f9c5df" containerName="glance-httpd" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.518698 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f34d682-a110-4d78-900b-da0ac3f9c5df" containerName="glance-httpd" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.518891 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="089e3832-1bfa-4c6a-9747-994ec8c5ee36" containerName="glance-httpd" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.518901 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="089e3832-1bfa-4c6a-9747-994ec8c5ee36" containerName="glance-log" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.518918 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad2af1c5-49df-4093-a38f-b1b47e2ddc1f" containerName="neutron-db-sync" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.518934 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f34d682-a110-4d78-900b-da0ac3f9c5df" containerName="glance-httpd" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.518947 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa3c4b78-d545-4bf5-95cd-33ef6670e895" containerName="init" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.518955 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f34d682-a110-4d78-900b-da0ac3f9c5df" containerName="glance-log" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.518964 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="18ba7210-7560-48f6-b164-780a6c84ef5f" containerName="keystone-bootstrap" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.519928 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.525438 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.526055 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.533936 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7ccdb5d4d7-59qkv"] Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.535399 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.536671 4559 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.536693 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.561010 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.567603 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/089e3832-1bfa-4c6a-9747-994ec8c5ee36-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "089e3832-1bfa-4c6a-9747-994ec8c5ee36" (UID: "089e3832-1bfa-4c6a-9747-994ec8c5ee36"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.577856 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/089e3832-1bfa-4c6a-9747-994ec8c5ee36-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "089e3832-1bfa-4c6a-9747-994ec8c5ee36" (UID: "089e3832-1bfa-4c6a-9747-994ec8c5ee36"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.580120 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ccdb5d4d7-59qkv"] Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.586635 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-67f8477cdd-4x7mh"] Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.588627 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "18ba7210-7560-48f6-b164-780a6c84ef5f" (UID: "18ba7210-7560-48f6-b164-780a6c84ef5f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.589718 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-67f8477cdd-4x7mh" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.592227 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-bqc2b" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.592449 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.592580 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.594099 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.604060 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-67f8477cdd-4x7mh"] Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.607694 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/089e3832-1bfa-4c6a-9747-994ec8c5ee36-config-data" (OuterVolumeSpecName: "config-data") pod "089e3832-1bfa-4c6a-9747-994ec8c5ee36" (UID: "089e3832-1bfa-4c6a-9747-994ec8c5ee36"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.637834 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.637898 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-logs\") pod \"glance-default-external-api-0\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.637927 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-dns-svc\") pod \"dnsmasq-dns-7ccdb5d4d7-59qkv\" (UID: \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\") " pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.637959 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-config\") pod \"dnsmasq-dns-7ccdb5d4d7-59qkv\" (UID: \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\") " pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.637985 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p42gk\" (UniqueName: \"kubernetes.io/projected/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-kube-api-access-p42gk\") pod \"dnsmasq-dns-7ccdb5d4d7-59qkv\" (UID: \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\") " pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.638008 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-config-data\") pod \"glance-default-external-api-0\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.638048 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.638066 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-ovsdbserver-sb\") pod \"dnsmasq-dns-7ccdb5d4d7-59qkv\" (UID: \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\") " pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.638117 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-ovsdbserver-nb\") pod \"dnsmasq-dns-7ccdb5d4d7-59qkv\" (UID: \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\") " pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.638203 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-scripts\") pod \"glance-default-external-api-0\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.638237 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8n5d\" (UniqueName: \"kubernetes.io/projected/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-kube-api-access-b8n5d\") pod \"glance-default-external-api-0\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.638266 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-dns-swift-storage-0\") pod \"dnsmasq-dns-7ccdb5d4d7-59qkv\" (UID: \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\") " pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.638281 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.638295 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.638355 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18ba7210-7560-48f6-b164-780a6c84ef5f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.638368 4559 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/089e3832-1bfa-4c6a-9747-994ec8c5ee36-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.638378 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/089e3832-1bfa-4c6a-9747-994ec8c5ee36-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.638385 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/089e3832-1bfa-4c6a-9747-994ec8c5ee36-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.721782 4559 scope.go:117] "RemoveContainer" containerID="c2ea7bf31dca742699dbe147d3090a60a159589e90db40c0601b82553039293b" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.740105 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-ovsdbserver-nb\") pod \"dnsmasq-dns-7ccdb5d4d7-59qkv\" (UID: \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\") " pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.740201 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4f62\" (UniqueName: \"kubernetes.io/projected/b649ece4-50a8-4449-ba98-2815000866fa-kube-api-access-t4f62\") pod \"neutron-67f8477cdd-4x7mh\" (UID: \"b649ece4-50a8-4449-ba98-2815000866fa\") " pod="openstack/neutron-67f8477cdd-4x7mh" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.740242 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b649ece4-50a8-4449-ba98-2815000866fa-ovndb-tls-certs\") pod \"neutron-67f8477cdd-4x7mh\" (UID: \"b649ece4-50a8-4449-ba98-2815000866fa\") " pod="openstack/neutron-67f8477cdd-4x7mh" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.740274 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-scripts\") pod \"glance-default-external-api-0\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.740309 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b649ece4-50a8-4449-ba98-2815000866fa-config\") pod \"neutron-67f8477cdd-4x7mh\" (UID: \"b649ece4-50a8-4449-ba98-2815000866fa\") " pod="openstack/neutron-67f8477cdd-4x7mh" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.740331 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8n5d\" (UniqueName: \"kubernetes.io/projected/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-kube-api-access-b8n5d\") pod \"glance-default-external-api-0\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.740363 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-dns-swift-storage-0\") pod \"dnsmasq-dns-7ccdb5d4d7-59qkv\" (UID: \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\") " pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.740382 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.740396 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.740443 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.740484 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-logs\") pod \"glance-default-external-api-0\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.740503 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-dns-svc\") pod \"dnsmasq-dns-7ccdb5d4d7-59qkv\" (UID: \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\") " pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.740532 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-config\") pod \"dnsmasq-dns-7ccdb5d4d7-59qkv\" (UID: \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\") " pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.740563 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p42gk\" (UniqueName: \"kubernetes.io/projected/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-kube-api-access-p42gk\") pod \"dnsmasq-dns-7ccdb5d4d7-59qkv\" (UID: \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\") " pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.740606 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-config-data\") pod \"glance-default-external-api-0\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.740663 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b649ece4-50a8-4449-ba98-2815000866fa-combined-ca-bundle\") pod \"neutron-67f8477cdd-4x7mh\" (UID: \"b649ece4-50a8-4449-ba98-2815000866fa\") " pod="openstack/neutron-67f8477cdd-4x7mh" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.740687 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.740711 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-ovsdbserver-sb\") pod \"dnsmasq-dns-7ccdb5d4d7-59qkv\" (UID: \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\") " pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.740747 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b649ece4-50a8-4449-ba98-2815000866fa-httpd-config\") pod \"neutron-67f8477cdd-4x7mh\" (UID: \"b649ece4-50a8-4449-ba98-2815000866fa\") " pod="openstack/neutron-67f8477cdd-4x7mh" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.741503 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.743518 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-dns-swift-storage-0\") pod \"dnsmasq-dns-7ccdb5d4d7-59qkv\" (UID: \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\") " pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.745593 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-logs\") pod \"glance-default-external-api-0\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.746936 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-config\") pod \"dnsmasq-dns-7ccdb5d4d7-59qkv\" (UID: \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\") " pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.747160 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-dns-svc\") pod \"dnsmasq-dns-7ccdb5d4d7-59qkv\" (UID: \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\") " pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.747968 4559 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-external-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.748832 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-ovsdbserver-sb\") pod \"dnsmasq-dns-7ccdb5d4d7-59qkv\" (UID: \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\") " pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.749207 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-648ff47655-pcz6l" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.750582 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-ovsdbserver-nb\") pod \"dnsmasq-dns-7ccdb5d4d7-59qkv\" (UID: \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\") " pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.753176 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.754846 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.756065 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-scripts\") pod \"glance-default-external-api-0\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.759486 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-config-data\") pod \"glance-default-external-api-0\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.766295 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p42gk\" (UniqueName: \"kubernetes.io/projected/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-kube-api-access-p42gk\") pod \"dnsmasq-dns-7ccdb5d4d7-59qkv\" (UID: \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\") " pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.780000 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8n5d\" (UniqueName: \"kubernetes.io/projected/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-kube-api-access-b8n5d\") pod \"glance-default-external-api-0\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.786751 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " pod="openstack/glance-default-external-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.841550 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-dns-swift-storage-0\") pod \"9dafbd81-21bc-4d7e-a61d-ba694502490c\" (UID: \"9dafbd81-21bc-4d7e-a61d-ba694502490c\") " Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.841604 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rj6rh\" (UniqueName: \"kubernetes.io/projected/9dafbd81-21bc-4d7e-a61d-ba694502490c-kube-api-access-rj6rh\") pod \"9dafbd81-21bc-4d7e-a61d-ba694502490c\" (UID: \"9dafbd81-21bc-4d7e-a61d-ba694502490c\") " Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.841667 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-ovsdbserver-nb\") pod \"9dafbd81-21bc-4d7e-a61d-ba694502490c\" (UID: \"9dafbd81-21bc-4d7e-a61d-ba694502490c\") " Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.841695 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-dns-svc\") pod \"9dafbd81-21bc-4d7e-a61d-ba694502490c\" (UID: \"9dafbd81-21bc-4d7e-a61d-ba694502490c\") " Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.841755 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-ovsdbserver-sb\") pod \"9dafbd81-21bc-4d7e-a61d-ba694502490c\" (UID: \"9dafbd81-21bc-4d7e-a61d-ba694502490c\") " Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.841868 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-config\") pod \"9dafbd81-21bc-4d7e-a61d-ba694502490c\" (UID: \"9dafbd81-21bc-4d7e-a61d-ba694502490c\") " Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.842184 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b649ece4-50a8-4449-ba98-2815000866fa-combined-ca-bundle\") pod \"neutron-67f8477cdd-4x7mh\" (UID: \"b649ece4-50a8-4449-ba98-2815000866fa\") " pod="openstack/neutron-67f8477cdd-4x7mh" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.842213 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b649ece4-50a8-4449-ba98-2815000866fa-httpd-config\") pod \"neutron-67f8477cdd-4x7mh\" (UID: \"b649ece4-50a8-4449-ba98-2815000866fa\") " pod="openstack/neutron-67f8477cdd-4x7mh" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.842277 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4f62\" (UniqueName: \"kubernetes.io/projected/b649ece4-50a8-4449-ba98-2815000866fa-kube-api-access-t4f62\") pod \"neutron-67f8477cdd-4x7mh\" (UID: \"b649ece4-50a8-4449-ba98-2815000866fa\") " pod="openstack/neutron-67f8477cdd-4x7mh" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.842301 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b649ece4-50a8-4449-ba98-2815000866fa-ovndb-tls-certs\") pod \"neutron-67f8477cdd-4x7mh\" (UID: \"b649ece4-50a8-4449-ba98-2815000866fa\") " pod="openstack/neutron-67f8477cdd-4x7mh" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.842337 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b649ece4-50a8-4449-ba98-2815000866fa-config\") pod \"neutron-67f8477cdd-4x7mh\" (UID: \"b649ece4-50a8-4449-ba98-2815000866fa\") " pod="openstack/neutron-67f8477cdd-4x7mh" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.848237 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b649ece4-50a8-4449-ba98-2815000866fa-combined-ca-bundle\") pod \"neutron-67f8477cdd-4x7mh\" (UID: \"b649ece4-50a8-4449-ba98-2815000866fa\") " pod="openstack/neutron-67f8477cdd-4x7mh" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.850195 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b649ece4-50a8-4449-ba98-2815000866fa-config\") pod \"neutron-67f8477cdd-4x7mh\" (UID: \"b649ece4-50a8-4449-ba98-2815000866fa\") " pod="openstack/neutron-67f8477cdd-4x7mh" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.851524 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b649ece4-50a8-4449-ba98-2815000866fa-httpd-config\") pod \"neutron-67f8477cdd-4x7mh\" (UID: \"b649ece4-50a8-4449-ba98-2815000866fa\") " pod="openstack/neutron-67f8477cdd-4x7mh" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.851537 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b649ece4-50a8-4449-ba98-2815000866fa-ovndb-tls-certs\") pod \"neutron-67f8477cdd-4x7mh\" (UID: \"b649ece4-50a8-4449-ba98-2815000866fa\") " pod="openstack/neutron-67f8477cdd-4x7mh" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.853809 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dafbd81-21bc-4d7e-a61d-ba694502490c-kube-api-access-rj6rh" (OuterVolumeSpecName: "kube-api-access-rj6rh") pod "9dafbd81-21bc-4d7e-a61d-ba694502490c" (UID: "9dafbd81-21bc-4d7e-a61d-ba694502490c"). InnerVolumeSpecName "kube-api-access-rj6rh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.862588 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4f62\" (UniqueName: \"kubernetes.io/projected/b649ece4-50a8-4449-ba98-2815000866fa-kube-api-access-t4f62\") pod \"neutron-67f8477cdd-4x7mh\" (UID: \"b649ece4-50a8-4449-ba98-2815000866fa\") " pod="openstack/neutron-67f8477cdd-4x7mh" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.911442 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.911753 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-67f8477cdd-4x7mh" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.912573 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.924941 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.934413 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9dafbd81-21bc-4d7e-a61d-ba694502490c" (UID: "9dafbd81-21bc-4d7e-a61d-ba694502490c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.934491 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 06:59:36 crc kubenswrapper[4559]: E1123 06:59:36.934935 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dafbd81-21bc-4d7e-a61d-ba694502490c" containerName="init" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.934949 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dafbd81-21bc-4d7e-a61d-ba694502490c" containerName="init" Nov 23 06:59:36 crc kubenswrapper[4559]: E1123 06:59:36.934964 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dafbd81-21bc-4d7e-a61d-ba694502490c" containerName="dnsmasq-dns" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.934970 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dafbd81-21bc-4d7e-a61d-ba694502490c" containerName="dnsmasq-dns" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.935173 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dafbd81-21bc-4d7e-a61d-ba694502490c" containerName="dnsmasq-dns" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.936096 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.937605 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.938819 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.939467 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.940095 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9dafbd81-21bc-4d7e-a61d-ba694502490c" (UID: "9dafbd81-21bc-4d7e-a61d-ba694502490c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.944754 4559 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.944785 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rj6rh\" (UniqueName: \"kubernetes.io/projected/9dafbd81-21bc-4d7e-a61d-ba694502490c-kube-api-access-rj6rh\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.944795 4559 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.948498 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-config" (OuterVolumeSpecName: "config") pod "9dafbd81-21bc-4d7e-a61d-ba694502490c" (UID: "9dafbd81-21bc-4d7e-a61d-ba694502490c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.951478 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9dafbd81-21bc-4d7e-a61d-ba694502490c" (UID: "9dafbd81-21bc-4d7e-a61d-ba694502490c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:36 crc kubenswrapper[4559]: I1123 06:59:36.959348 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9dafbd81-21bc-4d7e-a61d-ba694502490c" (UID: "9dafbd81-21bc-4d7e-a61d-ba694502490c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.038682 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.046173 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf03ed6e-e623-4ee8-afd4-361d6d18097a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.046257 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf03ed6e-e623-4ee8-afd4-361d6d18097a-logs\") pod \"glance-default-internal-api-0\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.046287 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf03ed6e-e623-4ee8-afd4-361d6d18097a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.046319 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf03ed6e-e623-4ee8-afd4-361d6d18097a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.046384 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bf03ed6e-e623-4ee8-afd4-361d6d18097a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.046543 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf03ed6e-e623-4ee8-afd4-361d6d18097a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.046568 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.046613 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f455s\" (UniqueName: \"kubernetes.io/projected/bf03ed6e-e623-4ee8-afd4-361d6d18097a-kube-api-access-f455s\") pod \"glance-default-internal-api-0\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.046686 4559 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.046697 4559 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.046706 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9dafbd81-21bc-4d7e-a61d-ba694502490c-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.147612 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bf03ed6e-e623-4ee8-afd4-361d6d18097a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.147776 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf03ed6e-e623-4ee8-afd4-361d6d18097a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.147809 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.147839 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f455s\" (UniqueName: \"kubernetes.io/projected/bf03ed6e-e623-4ee8-afd4-361d6d18097a-kube-api-access-f455s\") pod \"glance-default-internal-api-0\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.147874 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf03ed6e-e623-4ee8-afd4-361d6d18097a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.147917 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf03ed6e-e623-4ee8-afd4-361d6d18097a-logs\") pod \"glance-default-internal-api-0\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.147940 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf03ed6e-e623-4ee8-afd4-361d6d18097a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.147964 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf03ed6e-e623-4ee8-afd4-361d6d18097a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.148223 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bf03ed6e-e623-4ee8-afd4-361d6d18097a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.148429 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf03ed6e-e623-4ee8-afd4-361d6d18097a-logs\") pod \"glance-default-internal-api-0\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.148695 4559 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.152144 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf03ed6e-e623-4ee8-afd4-361d6d18097a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.153804 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf03ed6e-e623-4ee8-afd4-361d6d18097a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.166426 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf03ed6e-e623-4ee8-afd4-361d6d18097a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.168493 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf03ed6e-e623-4ee8-afd4-361d6d18097a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.169503 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f455s\" (UniqueName: \"kubernetes.io/projected/bf03ed6e-e623-4ee8-afd4-361d6d18097a-kube-api-access-f455s\") pod \"glance-default-internal-api-0\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.194571 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " pod="openstack/glance-default-internal-api-0" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.203577 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-nm8sn"] Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.211180 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-nm8sn"] Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.250769 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.299590 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-hz99c"] Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.302238 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hz99c" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.304799 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.304963 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.306035 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.306224 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.307242 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xmhpd" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.319358 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-hz99c"] Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.334903 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-k6zjg" event={"ID":"b14d2018-932c-47b6-bba9-232dff3ce320","Type":"ContainerStarted","Data":"cb75c9c2a61182d08852cade52ba64d0dd93308a8c9813f2b07c5c968f748421"} Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.350443 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-mvljp" event={"ID":"43aed302-5f95-4e51-b845-5a9cee9d1a08","Type":"ContainerStarted","Data":"112bc0f4b3c11d7acaad024495d518c9a4b4f11fca20992ff67b112146bba753"} Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.352051 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-k6zjg" podStartSLOduration=3.127042665 podStartE2EDuration="12.352034997s" podCreationTimestamp="2025-11-23 06:59:25 +0000 UTC" firstStartedPulling="2025-11-23 06:59:26.877209737 +0000 UTC m=+868.899195352" lastFinishedPulling="2025-11-23 06:59:36.10220207 +0000 UTC m=+878.124187684" observedRunningTime="2025-11-23 06:59:37.348371307 +0000 UTC m=+879.370356921" watchObservedRunningTime="2025-11-23 06:59:37.352034997 +0000 UTC m=+879.374020611" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.352290 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-scripts\") pod \"keystone-bootstrap-hz99c\" (UID: \"0534e805-e3fd-4193-80ed-3c9fa69a041a\") " pod="openstack/keystone-bootstrap-hz99c" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.352326 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-fernet-keys\") pod \"keystone-bootstrap-hz99c\" (UID: \"0534e805-e3fd-4193-80ed-3c9fa69a041a\") " pod="openstack/keystone-bootstrap-hz99c" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.352387 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-credential-keys\") pod \"keystone-bootstrap-hz99c\" (UID: \"0534e805-e3fd-4193-80ed-3c9fa69a041a\") " pod="openstack/keystone-bootstrap-hz99c" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.352401 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-combined-ca-bundle\") pod \"keystone-bootstrap-hz99c\" (UID: \"0534e805-e3fd-4193-80ed-3c9fa69a041a\") " pod="openstack/keystone-bootstrap-hz99c" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.352464 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-config-data\") pod \"keystone-bootstrap-hz99c\" (UID: \"0534e805-e3fd-4193-80ed-3c9fa69a041a\") " pod="openstack/keystone-bootstrap-hz99c" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.352493 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45ls6\" (UniqueName: \"kubernetes.io/projected/0534e805-e3fd-4193-80ed-3c9fa69a041a-kube-api-access-45ls6\") pod \"keystone-bootstrap-hz99c\" (UID: \"0534e805-e3fd-4193-80ed-3c9fa69a041a\") " pod="openstack/keystone-bootstrap-hz99c" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.376028 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-mvljp" podStartSLOduration=2.918691965 podStartE2EDuration="12.376006821s" podCreationTimestamp="2025-11-23 06:59:25 +0000 UTC" firstStartedPulling="2025-11-23 06:59:26.604965343 +0000 UTC m=+868.626950956" lastFinishedPulling="2025-11-23 06:59:36.062280198 +0000 UTC m=+878.084265812" observedRunningTime="2025-11-23 06:59:37.362920961 +0000 UTC m=+879.384906575" watchObservedRunningTime="2025-11-23 06:59:37.376006821 +0000 UTC m=+879.397992435" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.400333 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-648ff47655-pcz6l" event={"ID":"9dafbd81-21bc-4d7e-a61d-ba694502490c","Type":"ContainerDied","Data":"ed650f7549b264b84993bb17a274f71094c4cdd169ff3c9f95b62f50a41dd744"} Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.400559 4559 scope.go:117] "RemoveContainer" containerID="b87622aa82b32ff182c836aa8a7f812dd11a95f19917e11e867c5184307dc0c6" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.400831 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-648ff47655-pcz6l" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.406930 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4163bfe8-0de7-4490-b825-55d0f0343787","Type":"ContainerStarted","Data":"ba10afde1644658d448c1bb320ea05c40fdfb8995a31b98a833593e1267bbfbe"} Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.427260 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ccdb5d4d7-59qkv"] Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.455981 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-config-data\") pod \"keystone-bootstrap-hz99c\" (UID: \"0534e805-e3fd-4193-80ed-3c9fa69a041a\") " pod="openstack/keystone-bootstrap-hz99c" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.456046 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45ls6\" (UniqueName: \"kubernetes.io/projected/0534e805-e3fd-4193-80ed-3c9fa69a041a-kube-api-access-45ls6\") pod \"keystone-bootstrap-hz99c\" (UID: \"0534e805-e3fd-4193-80ed-3c9fa69a041a\") " pod="openstack/keystone-bootstrap-hz99c" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.456141 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-scripts\") pod \"keystone-bootstrap-hz99c\" (UID: \"0534e805-e3fd-4193-80ed-3c9fa69a041a\") " pod="openstack/keystone-bootstrap-hz99c" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.456202 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-fernet-keys\") pod \"keystone-bootstrap-hz99c\" (UID: \"0534e805-e3fd-4193-80ed-3c9fa69a041a\") " pod="openstack/keystone-bootstrap-hz99c" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.456307 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-credential-keys\") pod \"keystone-bootstrap-hz99c\" (UID: \"0534e805-e3fd-4193-80ed-3c9fa69a041a\") " pod="openstack/keystone-bootstrap-hz99c" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.456327 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-combined-ca-bundle\") pod \"keystone-bootstrap-hz99c\" (UID: \"0534e805-e3fd-4193-80ed-3c9fa69a041a\") " pod="openstack/keystone-bootstrap-hz99c" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.463655 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-config-data\") pod \"keystone-bootstrap-hz99c\" (UID: \"0534e805-e3fd-4193-80ed-3c9fa69a041a\") " pod="openstack/keystone-bootstrap-hz99c" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.466634 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-credential-keys\") pod \"keystone-bootstrap-hz99c\" (UID: \"0534e805-e3fd-4193-80ed-3c9fa69a041a\") " pod="openstack/keystone-bootstrap-hz99c" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.471051 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-scripts\") pod \"keystone-bootstrap-hz99c\" (UID: \"0534e805-e3fd-4193-80ed-3c9fa69a041a\") " pod="openstack/keystone-bootstrap-hz99c" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.471586 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-fernet-keys\") pod \"keystone-bootstrap-hz99c\" (UID: \"0534e805-e3fd-4193-80ed-3c9fa69a041a\") " pod="openstack/keystone-bootstrap-hz99c" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.475564 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-combined-ca-bundle\") pod \"keystone-bootstrap-hz99c\" (UID: \"0534e805-e3fd-4193-80ed-3c9fa69a041a\") " pod="openstack/keystone-bootstrap-hz99c" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.481165 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45ls6\" (UniqueName: \"kubernetes.io/projected/0534e805-e3fd-4193-80ed-3c9fa69a041a-kube-api-access-45ls6\") pod \"keystone-bootstrap-hz99c\" (UID: \"0534e805-e3fd-4193-80ed-3c9fa69a041a\") " pod="openstack/keystone-bootstrap-hz99c" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.521476 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-67f8477cdd-4x7mh"] Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.570069 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-648ff47655-pcz6l"] Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.576403 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-648ff47655-pcz6l"] Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.630410 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hz99c" Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.653769 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.705185 4559 scope.go:117] "RemoveContainer" containerID="1978a78cbcd124df30e44e33995c35b8b2834890d4a10ee9c29996cf7b34b370" Nov 23 06:59:37 crc kubenswrapper[4559]: W1123 06:59:37.712621 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3a4ae67_888b_46c2_bad1_92758aa2ea4f.slice/crio-7c7c7c5fadb1108aa464a6cd385b659fc3953e6743d84fe1112bb7ec1fa288c9 WatchSource:0}: Error finding container 7c7c7c5fadb1108aa464a6cd385b659fc3953e6743d84fe1112bb7ec1fa288c9: Status 404 returned error can't find the container with id 7c7c7c5fadb1108aa464a6cd385b659fc3953e6743d84fe1112bb7ec1fa288c9 Nov 23 06:59:37 crc kubenswrapper[4559]: I1123 06:59:37.912111 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 06:59:37 crc kubenswrapper[4559]: W1123 06:59:37.979192 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf03ed6e_e623_4ee8_afd4_361d6d18097a.slice/crio-7f6109ea8478ed9d7e5550b148afb17ec76a652e3d7aa331882555826881c2c9 WatchSource:0}: Error finding container 7f6109ea8478ed9d7e5550b148afb17ec76a652e3d7aa331882555826881c2c9: Status 404 returned error can't find the container with id 7f6109ea8478ed9d7e5550b148afb17ec76a652e3d7aa331882555826881c2c9 Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.180234 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-hz99c"] Nov 23 06:59:38 crc kubenswrapper[4559]: W1123 06:59:38.187412 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0534e805_e3fd_4193_80ed_3c9fa69a041a.slice/crio-e686b770243e7c35f771f1e539998d20e07bfbb18e5fb71c6af34d14c54cf1cc WatchSource:0}: Error finding container e686b770243e7c35f771f1e539998d20e07bfbb18e5fb71c6af34d14c54cf1cc: Status 404 returned error can't find the container with id e686b770243e7c35f771f1e539998d20e07bfbb18e5fb71c6af34d14c54cf1cc Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.301977 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="089e3832-1bfa-4c6a-9747-994ec8c5ee36" path="/var/lib/kubelet/pods/089e3832-1bfa-4c6a-9747-994ec8c5ee36/volumes" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.303237 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18ba7210-7560-48f6-b164-780a6c84ef5f" path="/var/lib/kubelet/pods/18ba7210-7560-48f6-b164-780a6c84ef5f/volumes" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.304205 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f34d682-a110-4d78-900b-da0ac3f9c5df" path="/var/lib/kubelet/pods/3f34d682-a110-4d78-900b-da0ac3f9c5df/volumes" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.305689 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dafbd81-21bc-4d7e-a61d-ba694502490c" path="/var/lib/kubelet/pods/9dafbd81-21bc-4d7e-a61d-ba694502490c/volumes" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.418059 4559 generic.go:334] "Generic (PLEG): container finished" podID="43aed302-5f95-4e51-b845-5a9cee9d1a08" containerID="112bc0f4b3c11d7acaad024495d518c9a4b4f11fca20992ff67b112146bba753" exitCode=0 Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.418113 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-mvljp" event={"ID":"43aed302-5f95-4e51-b845-5a9cee9d1a08","Type":"ContainerDied","Data":"112bc0f4b3c11d7acaad024495d518c9a4b4f11fca20992ff67b112146bba753"} Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.432155 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bf03ed6e-e623-4ee8-afd4-361d6d18097a","Type":"ContainerStarted","Data":"7f6109ea8478ed9d7e5550b148afb17ec76a652e3d7aa331882555826881c2c9"} Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.434539 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c3a4ae67-888b-46c2-bad1-92758aa2ea4f","Type":"ContainerStarted","Data":"a87a2cb07397e7a692e575ea7ebc473172586952ba546154c48fc11d0195ff43"} Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.434582 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c3a4ae67-888b-46c2-bad1-92758aa2ea4f","Type":"ContainerStarted","Data":"7c7c7c5fadb1108aa464a6cd385b659fc3953e6743d84fe1112bb7ec1fa288c9"} Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.436077 4559 generic.go:334] "Generic (PLEG): container finished" podID="5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0" containerID="58991aecfc4b746c4c67f4eda4d6dd9ee6884882102e6c8e63641970744f2578" exitCode=0 Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.436135 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" event={"ID":"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0","Type":"ContainerDied","Data":"58991aecfc4b746c4c67f4eda4d6dd9ee6884882102e6c8e63641970744f2578"} Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.436157 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" event={"ID":"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0","Type":"ContainerStarted","Data":"6d4c8038b911c2531b576d54af5fae2b9791edcbda45bf79948e6f8dd712af35"} Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.439709 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hz99c" event={"ID":"0534e805-e3fd-4193-80ed-3c9fa69a041a","Type":"ContainerStarted","Data":"e686b770243e7c35f771f1e539998d20e07bfbb18e5fb71c6af34d14c54cf1cc"} Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.446255 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67f8477cdd-4x7mh" event={"ID":"b649ece4-50a8-4449-ba98-2815000866fa","Type":"ContainerStarted","Data":"70a6af06589557c859bf0b9b1efe59d4f712ca7df817304e61ba10633adc6a9b"} Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.446297 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-67f8477cdd-4x7mh" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.446308 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67f8477cdd-4x7mh" event={"ID":"b649ece4-50a8-4449-ba98-2815000866fa","Type":"ContainerStarted","Data":"bd5950999e0816dc3a6c6c59ae79a87d4896d498e6aec37ee3fb20cea46e18c3"} Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.446318 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67f8477cdd-4x7mh" event={"ID":"b649ece4-50a8-4449-ba98-2815000866fa","Type":"ContainerStarted","Data":"20e1977fa028bc9ca61faf1d37d02151b2d4a56fa46d18a3c1edf628df12f3a7"} Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.480978 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-67f8477cdd-4x7mh" podStartSLOduration=2.480957106 podStartE2EDuration="2.480957106s" podCreationTimestamp="2025-11-23 06:59:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:59:38.462830604 +0000 UTC m=+880.484816218" watchObservedRunningTime="2025-11-23 06:59:38.480957106 +0000 UTC m=+880.502942720" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.539614 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-d8c98f585-kr7vt"] Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.541792 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d8c98f585-kr7vt" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.546261 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.547022 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.568211 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d8c98f585-kr7vt"] Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.698695 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjjmg\" (UniqueName: \"kubernetes.io/projected/72027df5-f460-436a-b883-1895caea6f90-kube-api-access-vjjmg\") pod \"neutron-d8c98f585-kr7vt\" (UID: \"72027df5-f460-436a-b883-1895caea6f90\") " pod="openstack/neutron-d8c98f585-kr7vt" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.699212 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72027df5-f460-436a-b883-1895caea6f90-public-tls-certs\") pod \"neutron-d8c98f585-kr7vt\" (UID: \"72027df5-f460-436a-b883-1895caea6f90\") " pod="openstack/neutron-d8c98f585-kr7vt" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.699262 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/72027df5-f460-436a-b883-1895caea6f90-config\") pod \"neutron-d8c98f585-kr7vt\" (UID: \"72027df5-f460-436a-b883-1895caea6f90\") " pod="openstack/neutron-d8c98f585-kr7vt" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.699290 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72027df5-f460-436a-b883-1895caea6f90-combined-ca-bundle\") pod \"neutron-d8c98f585-kr7vt\" (UID: \"72027df5-f460-436a-b883-1895caea6f90\") " pod="openstack/neutron-d8c98f585-kr7vt" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.699408 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/72027df5-f460-436a-b883-1895caea6f90-internal-tls-certs\") pod \"neutron-d8c98f585-kr7vt\" (UID: \"72027df5-f460-436a-b883-1895caea6f90\") " pod="openstack/neutron-d8c98f585-kr7vt" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.699455 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/72027df5-f460-436a-b883-1895caea6f90-ovndb-tls-certs\") pod \"neutron-d8c98f585-kr7vt\" (UID: \"72027df5-f460-436a-b883-1895caea6f90\") " pod="openstack/neutron-d8c98f585-kr7vt" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.699479 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/72027df5-f460-436a-b883-1895caea6f90-httpd-config\") pod \"neutron-d8c98f585-kr7vt\" (UID: \"72027df5-f460-436a-b883-1895caea6f90\") " pod="openstack/neutron-d8c98f585-kr7vt" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.801187 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72027df5-f460-436a-b883-1895caea6f90-public-tls-certs\") pod \"neutron-d8c98f585-kr7vt\" (UID: \"72027df5-f460-436a-b883-1895caea6f90\") " pod="openstack/neutron-d8c98f585-kr7vt" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.801251 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/72027df5-f460-436a-b883-1895caea6f90-config\") pod \"neutron-d8c98f585-kr7vt\" (UID: \"72027df5-f460-436a-b883-1895caea6f90\") " pod="openstack/neutron-d8c98f585-kr7vt" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.801278 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72027df5-f460-436a-b883-1895caea6f90-combined-ca-bundle\") pod \"neutron-d8c98f585-kr7vt\" (UID: \"72027df5-f460-436a-b883-1895caea6f90\") " pod="openstack/neutron-d8c98f585-kr7vt" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.801367 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/72027df5-f460-436a-b883-1895caea6f90-internal-tls-certs\") pod \"neutron-d8c98f585-kr7vt\" (UID: \"72027df5-f460-436a-b883-1895caea6f90\") " pod="openstack/neutron-d8c98f585-kr7vt" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.802458 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/72027df5-f460-436a-b883-1895caea6f90-ovndb-tls-certs\") pod \"neutron-d8c98f585-kr7vt\" (UID: \"72027df5-f460-436a-b883-1895caea6f90\") " pod="openstack/neutron-d8c98f585-kr7vt" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.802517 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/72027df5-f460-436a-b883-1895caea6f90-httpd-config\") pod \"neutron-d8c98f585-kr7vt\" (UID: \"72027df5-f460-436a-b883-1895caea6f90\") " pod="openstack/neutron-d8c98f585-kr7vt" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.802604 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjjmg\" (UniqueName: \"kubernetes.io/projected/72027df5-f460-436a-b883-1895caea6f90-kube-api-access-vjjmg\") pod \"neutron-d8c98f585-kr7vt\" (UID: \"72027df5-f460-436a-b883-1895caea6f90\") " pod="openstack/neutron-d8c98f585-kr7vt" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.807370 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/72027df5-f460-436a-b883-1895caea6f90-internal-tls-certs\") pod \"neutron-d8c98f585-kr7vt\" (UID: \"72027df5-f460-436a-b883-1895caea6f90\") " pod="openstack/neutron-d8c98f585-kr7vt" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.807997 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72027df5-f460-436a-b883-1895caea6f90-public-tls-certs\") pod \"neutron-d8c98f585-kr7vt\" (UID: \"72027df5-f460-436a-b883-1895caea6f90\") " pod="openstack/neutron-d8c98f585-kr7vt" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.808860 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72027df5-f460-436a-b883-1895caea6f90-combined-ca-bundle\") pod \"neutron-d8c98f585-kr7vt\" (UID: \"72027df5-f460-436a-b883-1895caea6f90\") " pod="openstack/neutron-d8c98f585-kr7vt" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.809980 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/72027df5-f460-436a-b883-1895caea6f90-httpd-config\") pod \"neutron-d8c98f585-kr7vt\" (UID: \"72027df5-f460-436a-b883-1895caea6f90\") " pod="openstack/neutron-d8c98f585-kr7vt" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.810493 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/72027df5-f460-436a-b883-1895caea6f90-ovndb-tls-certs\") pod \"neutron-d8c98f585-kr7vt\" (UID: \"72027df5-f460-436a-b883-1895caea6f90\") " pod="openstack/neutron-d8c98f585-kr7vt" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.811820 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/72027df5-f460-436a-b883-1895caea6f90-config\") pod \"neutron-d8c98f585-kr7vt\" (UID: \"72027df5-f460-436a-b883-1895caea6f90\") " pod="openstack/neutron-d8c98f585-kr7vt" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.816883 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjjmg\" (UniqueName: \"kubernetes.io/projected/72027df5-f460-436a-b883-1895caea6f90-kube-api-access-vjjmg\") pod \"neutron-d8c98f585-kr7vt\" (UID: \"72027df5-f460-436a-b883-1895caea6f90\") " pod="openstack/neutron-d8c98f585-kr7vt" Nov 23 06:59:38 crc kubenswrapper[4559]: I1123 06:59:38.879356 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d8c98f585-kr7vt" Nov 23 06:59:39 crc kubenswrapper[4559]: I1123 06:59:39.386682 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d8c98f585-kr7vt"] Nov 23 06:59:39 crc kubenswrapper[4559]: W1123 06:59:39.404266 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod72027df5_f460_436a_b883_1895caea6f90.slice/crio-50896af8923051c0cacfeab2df9133398cd0188c639ca96cbaabf6c599d2e925 WatchSource:0}: Error finding container 50896af8923051c0cacfeab2df9133398cd0188c639ca96cbaabf6c599d2e925: Status 404 returned error can't find the container with id 50896af8923051c0cacfeab2df9133398cd0188c639ca96cbaabf6c599d2e925 Nov 23 06:59:39 crc kubenswrapper[4559]: I1123 06:59:39.457615 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4163bfe8-0de7-4490-b825-55d0f0343787","Type":"ContainerStarted","Data":"cd785a0026c59d5e74c27422561bc4c114f548239eeb6189968a2dd70cd1543d"} Nov 23 06:59:39 crc kubenswrapper[4559]: I1123 06:59:39.460771 4559 generic.go:334] "Generic (PLEG): container finished" podID="b14d2018-932c-47b6-bba9-232dff3ce320" containerID="cb75c9c2a61182d08852cade52ba64d0dd93308a8c9813f2b07c5c968f748421" exitCode=0 Nov 23 06:59:39 crc kubenswrapper[4559]: I1123 06:59:39.460825 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-k6zjg" event={"ID":"b14d2018-932c-47b6-bba9-232dff3ce320","Type":"ContainerDied","Data":"cb75c9c2a61182d08852cade52ba64d0dd93308a8c9813f2b07c5c968f748421"} Nov 23 06:59:39 crc kubenswrapper[4559]: I1123 06:59:39.463079 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bf03ed6e-e623-4ee8-afd4-361d6d18097a","Type":"ContainerStarted","Data":"a2503ed47292a8e094ec1cee0b8416e38dfefebe40c846cff1fd8c7958096c24"} Nov 23 06:59:39 crc kubenswrapper[4559]: I1123 06:59:39.467091 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" event={"ID":"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0","Type":"ContainerStarted","Data":"d12b26a69c747e43b76d9f3e19c15607447fb89167840665827bd36455f11a6a"} Nov 23 06:59:39 crc kubenswrapper[4559]: I1123 06:59:39.468051 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" Nov 23 06:59:39 crc kubenswrapper[4559]: I1123 06:59:39.469155 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hz99c" event={"ID":"0534e805-e3fd-4193-80ed-3c9fa69a041a","Type":"ContainerStarted","Data":"ffc5cf05f27d9f9aba701eab360263ff3bebe04d77619e8fa4aded2a481125eb"} Nov 23 06:59:39 crc kubenswrapper[4559]: I1123 06:59:39.483003 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d8c98f585-kr7vt" event={"ID":"72027df5-f460-436a-b883-1895caea6f90","Type":"ContainerStarted","Data":"50896af8923051c0cacfeab2df9133398cd0188c639ca96cbaabf6c599d2e925"} Nov 23 06:59:39 crc kubenswrapper[4559]: I1123 06:59:39.504686 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" podStartSLOduration=3.504666477 podStartE2EDuration="3.504666477s" podCreationTimestamp="2025-11-23 06:59:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:59:39.497390463 +0000 UTC m=+881.519376077" watchObservedRunningTime="2025-11-23 06:59:39.504666477 +0000 UTC m=+881.526652091" Nov 23 06:59:39 crc kubenswrapper[4559]: I1123 06:59:39.516296 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-hz99c" podStartSLOduration=2.516275511 podStartE2EDuration="2.516275511s" podCreationTimestamp="2025-11-23 06:59:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:59:39.508357399 +0000 UTC m=+881.530343013" watchObservedRunningTime="2025-11-23 06:59:39.516275511 +0000 UTC m=+881.538261125" Nov 23 06:59:39 crc kubenswrapper[4559]: I1123 06:59:39.734493 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-mvljp" Nov 23 06:59:39 crc kubenswrapper[4559]: I1123 06:59:39.825689 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43aed302-5f95-4e51-b845-5a9cee9d1a08-config-data\") pod \"43aed302-5f95-4e51-b845-5a9cee9d1a08\" (UID: \"43aed302-5f95-4e51-b845-5a9cee9d1a08\") " Nov 23 06:59:39 crc kubenswrapper[4559]: I1123 06:59:39.825744 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43aed302-5f95-4e51-b845-5a9cee9d1a08-logs\") pod \"43aed302-5f95-4e51-b845-5a9cee9d1a08\" (UID: \"43aed302-5f95-4e51-b845-5a9cee9d1a08\") " Nov 23 06:59:39 crc kubenswrapper[4559]: I1123 06:59:39.825822 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43aed302-5f95-4e51-b845-5a9cee9d1a08-scripts\") pod \"43aed302-5f95-4e51-b845-5a9cee9d1a08\" (UID: \"43aed302-5f95-4e51-b845-5a9cee9d1a08\") " Nov 23 06:59:39 crc kubenswrapper[4559]: I1123 06:59:39.825884 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43aed302-5f95-4e51-b845-5a9cee9d1a08-combined-ca-bundle\") pod \"43aed302-5f95-4e51-b845-5a9cee9d1a08\" (UID: \"43aed302-5f95-4e51-b845-5a9cee9d1a08\") " Nov 23 06:59:39 crc kubenswrapper[4559]: I1123 06:59:39.825920 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k557r\" (UniqueName: \"kubernetes.io/projected/43aed302-5f95-4e51-b845-5a9cee9d1a08-kube-api-access-k557r\") pod \"43aed302-5f95-4e51-b845-5a9cee9d1a08\" (UID: \"43aed302-5f95-4e51-b845-5a9cee9d1a08\") " Nov 23 06:59:39 crc kubenswrapper[4559]: I1123 06:59:39.827076 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43aed302-5f95-4e51-b845-5a9cee9d1a08-logs" (OuterVolumeSpecName: "logs") pod "43aed302-5f95-4e51-b845-5a9cee9d1a08" (UID: "43aed302-5f95-4e51-b845-5a9cee9d1a08"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 06:59:39 crc kubenswrapper[4559]: I1123 06:59:39.833787 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43aed302-5f95-4e51-b845-5a9cee9d1a08-scripts" (OuterVolumeSpecName: "scripts") pod "43aed302-5f95-4e51-b845-5a9cee9d1a08" (UID: "43aed302-5f95-4e51-b845-5a9cee9d1a08"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:39 crc kubenswrapper[4559]: I1123 06:59:39.834756 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43aed302-5f95-4e51-b845-5a9cee9d1a08-kube-api-access-k557r" (OuterVolumeSpecName: "kube-api-access-k557r") pod "43aed302-5f95-4e51-b845-5a9cee9d1a08" (UID: "43aed302-5f95-4e51-b845-5a9cee9d1a08"). InnerVolumeSpecName "kube-api-access-k557r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:59:39 crc kubenswrapper[4559]: I1123 06:59:39.877763 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43aed302-5f95-4e51-b845-5a9cee9d1a08-config-data" (OuterVolumeSpecName: "config-data") pod "43aed302-5f95-4e51-b845-5a9cee9d1a08" (UID: "43aed302-5f95-4e51-b845-5a9cee9d1a08"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:39 crc kubenswrapper[4559]: I1123 06:59:39.877854 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43aed302-5f95-4e51-b845-5a9cee9d1a08-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "43aed302-5f95-4e51-b845-5a9cee9d1a08" (UID: "43aed302-5f95-4e51-b845-5a9cee9d1a08"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:39 crc kubenswrapper[4559]: I1123 06:59:39.927719 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43aed302-5f95-4e51-b845-5a9cee9d1a08-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:39 crc kubenswrapper[4559]: I1123 06:59:39.927748 4559 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43aed302-5f95-4e51-b845-5a9cee9d1a08-logs\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:39 crc kubenswrapper[4559]: I1123 06:59:39.927758 4559 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43aed302-5f95-4e51-b845-5a9cee9d1a08-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:39 crc kubenswrapper[4559]: I1123 06:59:39.927768 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43aed302-5f95-4e51-b845-5a9cee9d1a08-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:39 crc kubenswrapper[4559]: I1123 06:59:39.927778 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k557r\" (UniqueName: \"kubernetes.io/projected/43aed302-5f95-4e51-b845-5a9cee9d1a08-kube-api-access-k557r\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.457624 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7667f44776-mszxt"] Nov 23 06:59:40 crc kubenswrapper[4559]: E1123 06:59:40.458341 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43aed302-5f95-4e51-b845-5a9cee9d1a08" containerName="placement-db-sync" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.458359 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="43aed302-5f95-4e51-b845-5a9cee9d1a08" containerName="placement-db-sync" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.458549 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="43aed302-5f95-4e51-b845-5a9cee9d1a08" containerName="placement-db-sync" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.459711 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7667f44776-mszxt" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.461846 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.461969 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.475984 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7667f44776-mszxt"] Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.497889 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-mvljp" event={"ID":"43aed302-5f95-4e51-b845-5a9cee9d1a08","Type":"ContainerDied","Data":"357376614a62e7b787fddd897553a74c3cb166227d16413f80afe94f71321ae9"} Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.497920 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-mvljp" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.497940 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="357376614a62e7b787fddd897553a74c3cb166227d16413f80afe94f71321ae9" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.500202 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bf03ed6e-e623-4ee8-afd4-361d6d18097a","Type":"ContainerStarted","Data":"cbd4d0e094e31047191103ed3e0e5d8a94e077ebe5cb31c5ec25c3b7850932a0"} Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.517939 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.517918442 podStartE2EDuration="4.517918442s" podCreationTimestamp="2025-11-23 06:59:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:59:40.516632114 +0000 UTC m=+882.538617727" watchObservedRunningTime="2025-11-23 06:59:40.517918442 +0000 UTC m=+882.539904056" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.519865 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c3a4ae67-888b-46c2-bad1-92758aa2ea4f","Type":"ContainerStarted","Data":"27f8bbb0ad0a072b9c09d2f2502d82cb90133957a9ee3c475747e68d2e5c9159"} Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.524474 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d8c98f585-kr7vt" event={"ID":"72027df5-f460-436a-b883-1895caea6f90","Type":"ContainerStarted","Data":"3cd6c6e345f80846a3b578d70e0de998f760c9298e540d0f28344298bb071325"} Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.524520 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d8c98f585-kr7vt" event={"ID":"72027df5-f460-436a-b883-1895caea6f90","Type":"ContainerStarted","Data":"1f2d07152e511fa6ff4e9673f60b312438948d9af56e7ea9e9e357141aa1ea2c"} Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.524743 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-d8c98f585-kr7vt" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.541611 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gq2fr\" (UniqueName: \"kubernetes.io/projected/1df68d96-dfae-42c9-8a11-9906486ea1b9-kube-api-access-gq2fr\") pod \"placement-7667f44776-mszxt\" (UID: \"1df68d96-dfae-42c9-8a11-9906486ea1b9\") " pod="openstack/placement-7667f44776-mszxt" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.541738 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1df68d96-dfae-42c9-8a11-9906486ea1b9-logs\") pod \"placement-7667f44776-mszxt\" (UID: \"1df68d96-dfae-42c9-8a11-9906486ea1b9\") " pod="openstack/placement-7667f44776-mszxt" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.541795 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1df68d96-dfae-42c9-8a11-9906486ea1b9-internal-tls-certs\") pod \"placement-7667f44776-mszxt\" (UID: \"1df68d96-dfae-42c9-8a11-9906486ea1b9\") " pod="openstack/placement-7667f44776-mszxt" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.542010 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1df68d96-dfae-42c9-8a11-9906486ea1b9-scripts\") pod \"placement-7667f44776-mszxt\" (UID: \"1df68d96-dfae-42c9-8a11-9906486ea1b9\") " pod="openstack/placement-7667f44776-mszxt" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.542090 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1df68d96-dfae-42c9-8a11-9906486ea1b9-combined-ca-bundle\") pod \"placement-7667f44776-mszxt\" (UID: \"1df68d96-dfae-42c9-8a11-9906486ea1b9\") " pod="openstack/placement-7667f44776-mszxt" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.542126 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1df68d96-dfae-42c9-8a11-9906486ea1b9-public-tls-certs\") pod \"placement-7667f44776-mszxt\" (UID: \"1df68d96-dfae-42c9-8a11-9906486ea1b9\") " pod="openstack/placement-7667f44776-mszxt" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.542192 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1df68d96-dfae-42c9-8a11-9906486ea1b9-config-data\") pod \"placement-7667f44776-mszxt\" (UID: \"1df68d96-dfae-42c9-8a11-9906486ea1b9\") " pod="openstack/placement-7667f44776-mszxt" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.543604 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.543587369 podStartE2EDuration="4.543587369s" podCreationTimestamp="2025-11-23 06:59:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:59:40.539544835 +0000 UTC m=+882.561530450" watchObservedRunningTime="2025-11-23 06:59:40.543587369 +0000 UTC m=+882.565572983" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.554840 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-d8c98f585-kr7vt" podStartSLOduration=2.5548001769999997 podStartE2EDuration="2.554800177s" podCreationTimestamp="2025-11-23 06:59:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:59:40.552242155 +0000 UTC m=+882.574227769" watchObservedRunningTime="2025-11-23 06:59:40.554800177 +0000 UTC m=+882.576785791" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.644531 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1df68d96-dfae-42c9-8a11-9906486ea1b9-internal-tls-certs\") pod \"placement-7667f44776-mszxt\" (UID: \"1df68d96-dfae-42c9-8a11-9906486ea1b9\") " pod="openstack/placement-7667f44776-mszxt" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.644614 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1df68d96-dfae-42c9-8a11-9906486ea1b9-scripts\") pod \"placement-7667f44776-mszxt\" (UID: \"1df68d96-dfae-42c9-8a11-9906486ea1b9\") " pod="openstack/placement-7667f44776-mszxt" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.644711 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1df68d96-dfae-42c9-8a11-9906486ea1b9-combined-ca-bundle\") pod \"placement-7667f44776-mszxt\" (UID: \"1df68d96-dfae-42c9-8a11-9906486ea1b9\") " pod="openstack/placement-7667f44776-mszxt" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.644763 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1df68d96-dfae-42c9-8a11-9906486ea1b9-public-tls-certs\") pod \"placement-7667f44776-mszxt\" (UID: \"1df68d96-dfae-42c9-8a11-9906486ea1b9\") " pod="openstack/placement-7667f44776-mszxt" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.644829 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1df68d96-dfae-42c9-8a11-9906486ea1b9-config-data\") pod \"placement-7667f44776-mszxt\" (UID: \"1df68d96-dfae-42c9-8a11-9906486ea1b9\") " pod="openstack/placement-7667f44776-mszxt" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.644871 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gq2fr\" (UniqueName: \"kubernetes.io/projected/1df68d96-dfae-42c9-8a11-9906486ea1b9-kube-api-access-gq2fr\") pod \"placement-7667f44776-mszxt\" (UID: \"1df68d96-dfae-42c9-8a11-9906486ea1b9\") " pod="openstack/placement-7667f44776-mszxt" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.645025 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1df68d96-dfae-42c9-8a11-9906486ea1b9-logs\") pod \"placement-7667f44776-mszxt\" (UID: \"1df68d96-dfae-42c9-8a11-9906486ea1b9\") " pod="openstack/placement-7667f44776-mszxt" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.645593 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1df68d96-dfae-42c9-8a11-9906486ea1b9-logs\") pod \"placement-7667f44776-mszxt\" (UID: \"1df68d96-dfae-42c9-8a11-9906486ea1b9\") " pod="openstack/placement-7667f44776-mszxt" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.653306 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1df68d96-dfae-42c9-8a11-9906486ea1b9-internal-tls-certs\") pod \"placement-7667f44776-mszxt\" (UID: \"1df68d96-dfae-42c9-8a11-9906486ea1b9\") " pod="openstack/placement-7667f44776-mszxt" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.653555 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1df68d96-dfae-42c9-8a11-9906486ea1b9-public-tls-certs\") pod \"placement-7667f44776-mszxt\" (UID: \"1df68d96-dfae-42c9-8a11-9906486ea1b9\") " pod="openstack/placement-7667f44776-mszxt" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.653988 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1df68d96-dfae-42c9-8a11-9906486ea1b9-config-data\") pod \"placement-7667f44776-mszxt\" (UID: \"1df68d96-dfae-42c9-8a11-9906486ea1b9\") " pod="openstack/placement-7667f44776-mszxt" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.661482 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gq2fr\" (UniqueName: \"kubernetes.io/projected/1df68d96-dfae-42c9-8a11-9906486ea1b9-kube-api-access-gq2fr\") pod \"placement-7667f44776-mszxt\" (UID: \"1df68d96-dfae-42c9-8a11-9906486ea1b9\") " pod="openstack/placement-7667f44776-mszxt" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.662718 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1df68d96-dfae-42c9-8a11-9906486ea1b9-scripts\") pod \"placement-7667f44776-mszxt\" (UID: \"1df68d96-dfae-42c9-8a11-9906486ea1b9\") " pod="openstack/placement-7667f44776-mszxt" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.671260 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1df68d96-dfae-42c9-8a11-9906486ea1b9-combined-ca-bundle\") pod \"placement-7667f44776-mszxt\" (UID: \"1df68d96-dfae-42c9-8a11-9906486ea1b9\") " pod="openstack/placement-7667f44776-mszxt" Nov 23 06:59:40 crc kubenswrapper[4559]: I1123 06:59:40.777588 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7667f44776-mszxt" Nov 23 06:59:41 crc kubenswrapper[4559]: I1123 06:59:41.538311 4559 generic.go:334] "Generic (PLEG): container finished" podID="0534e805-e3fd-4193-80ed-3c9fa69a041a" containerID="ffc5cf05f27d9f9aba701eab360263ff3bebe04d77619e8fa4aded2a481125eb" exitCode=0 Nov 23 06:59:41 crc kubenswrapper[4559]: I1123 06:59:41.538407 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hz99c" event={"ID":"0534e805-e3fd-4193-80ed-3c9fa69a041a","Type":"ContainerDied","Data":"ffc5cf05f27d9f9aba701eab360263ff3bebe04d77619e8fa4aded2a481125eb"} Nov 23 06:59:46 crc kubenswrapper[4559]: I1123 06:59:46.912827 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" Nov 23 06:59:46 crc kubenswrapper[4559]: I1123 06:59:46.952034 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d78ff46f5-kdzjr"] Nov 23 06:59:46 crc kubenswrapper[4559]: I1123 06:59:46.952300 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" podUID="58898cc0-18e6-41c6-9ce2-149d081ff29b" containerName="dnsmasq-dns" containerID="cri-o://93860435ee52af0972f94019211071574356bbc2034f1651ff6d61882b1e7bbe" gracePeriod=10 Nov 23 06:59:47 crc kubenswrapper[4559]: I1123 06:59:47.039765 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 23 06:59:47 crc kubenswrapper[4559]: I1123 06:59:47.039894 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 23 06:59:47 crc kubenswrapper[4559]: I1123 06:59:47.064807 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 23 06:59:47 crc kubenswrapper[4559]: I1123 06:59:47.068506 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 23 06:59:47 crc kubenswrapper[4559]: I1123 06:59:47.251978 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 23 06:59:47 crc kubenswrapper[4559]: I1123 06:59:47.252041 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 23 06:59:47 crc kubenswrapper[4559]: I1123 06:59:47.286417 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 23 06:59:47 crc kubenswrapper[4559]: I1123 06:59:47.299368 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 23 06:59:47 crc kubenswrapper[4559]: I1123 06:59:47.614104 4559 generic.go:334] "Generic (PLEG): container finished" podID="58898cc0-18e6-41c6-9ce2-149d081ff29b" containerID="93860435ee52af0972f94019211071574356bbc2034f1651ff6d61882b1e7bbe" exitCode=0 Nov 23 06:59:47 crc kubenswrapper[4559]: I1123 06:59:47.614176 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" event={"ID":"58898cc0-18e6-41c6-9ce2-149d081ff29b","Type":"ContainerDied","Data":"93860435ee52af0972f94019211071574356bbc2034f1651ff6d61882b1e7bbe"} Nov 23 06:59:47 crc kubenswrapper[4559]: I1123 06:59:47.614856 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 23 06:59:47 crc kubenswrapper[4559]: I1123 06:59:47.614906 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 23 06:59:47 crc kubenswrapper[4559]: I1123 06:59:47.614928 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 23 06:59:47 crc kubenswrapper[4559]: I1123 06:59:47.614943 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.377891 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-k6zjg" Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.386771 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hz99c" Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.403385 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-scripts\") pod \"0534e805-e3fd-4193-80ed-3c9fa69a041a\" (UID: \"0534e805-e3fd-4193-80ed-3c9fa69a041a\") " Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.403471 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b14d2018-932c-47b6-bba9-232dff3ce320-db-sync-config-data\") pod \"b14d2018-932c-47b6-bba9-232dff3ce320\" (UID: \"b14d2018-932c-47b6-bba9-232dff3ce320\") " Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.403546 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b14d2018-932c-47b6-bba9-232dff3ce320-combined-ca-bundle\") pod \"b14d2018-932c-47b6-bba9-232dff3ce320\" (UID: \"b14d2018-932c-47b6-bba9-232dff3ce320\") " Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.403670 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-fernet-keys\") pod \"0534e805-e3fd-4193-80ed-3c9fa69a041a\" (UID: \"0534e805-e3fd-4193-80ed-3c9fa69a041a\") " Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.403701 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-combined-ca-bundle\") pod \"0534e805-e3fd-4193-80ed-3c9fa69a041a\" (UID: \"0534e805-e3fd-4193-80ed-3c9fa69a041a\") " Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.403737 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45ls6\" (UniqueName: \"kubernetes.io/projected/0534e805-e3fd-4193-80ed-3c9fa69a041a-kube-api-access-45ls6\") pod \"0534e805-e3fd-4193-80ed-3c9fa69a041a\" (UID: \"0534e805-e3fd-4193-80ed-3c9fa69a041a\") " Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.403836 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jn9c7\" (UniqueName: \"kubernetes.io/projected/b14d2018-932c-47b6-bba9-232dff3ce320-kube-api-access-jn9c7\") pod \"b14d2018-932c-47b6-bba9-232dff3ce320\" (UID: \"b14d2018-932c-47b6-bba9-232dff3ce320\") " Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.403882 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-credential-keys\") pod \"0534e805-e3fd-4193-80ed-3c9fa69a041a\" (UID: \"0534e805-e3fd-4193-80ed-3c9fa69a041a\") " Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.403939 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-config-data\") pod \"0534e805-e3fd-4193-80ed-3c9fa69a041a\" (UID: \"0534e805-e3fd-4193-80ed-3c9fa69a041a\") " Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.427017 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "0534e805-e3fd-4193-80ed-3c9fa69a041a" (UID: "0534e805-e3fd-4193-80ed-3c9fa69a041a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.427751 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0534e805-e3fd-4193-80ed-3c9fa69a041a-kube-api-access-45ls6" (OuterVolumeSpecName: "kube-api-access-45ls6") pod "0534e805-e3fd-4193-80ed-3c9fa69a041a" (UID: "0534e805-e3fd-4193-80ed-3c9fa69a041a"). InnerVolumeSpecName "kube-api-access-45ls6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.436206 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-scripts" (OuterVolumeSpecName: "scripts") pod "0534e805-e3fd-4193-80ed-3c9fa69a041a" (UID: "0534e805-e3fd-4193-80ed-3c9fa69a041a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.436295 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b14d2018-932c-47b6-bba9-232dff3ce320-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "b14d2018-932c-47b6-bba9-232dff3ce320" (UID: "b14d2018-932c-47b6-bba9-232dff3ce320"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.444849 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b14d2018-932c-47b6-bba9-232dff3ce320-kube-api-access-jn9c7" (OuterVolumeSpecName: "kube-api-access-jn9c7") pod "b14d2018-932c-47b6-bba9-232dff3ce320" (UID: "b14d2018-932c-47b6-bba9-232dff3ce320"). InnerVolumeSpecName "kube-api-access-jn9c7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.456766 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "0534e805-e3fd-4193-80ed-3c9fa69a041a" (UID: "0534e805-e3fd-4193-80ed-3c9fa69a041a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.477186 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b14d2018-932c-47b6-bba9-232dff3ce320-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b14d2018-932c-47b6-bba9-232dff3ce320" (UID: "b14d2018-932c-47b6-bba9-232dff3ce320"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.478900 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0534e805-e3fd-4193-80ed-3c9fa69a041a" (UID: "0534e805-e3fd-4193-80ed-3c9fa69a041a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.482715 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-config-data" (OuterVolumeSpecName: "config-data") pod "0534e805-e3fd-4193-80ed-3c9fa69a041a" (UID: "0534e805-e3fd-4193-80ed-3c9fa69a041a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.505921 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jn9c7\" (UniqueName: \"kubernetes.io/projected/b14d2018-932c-47b6-bba9-232dff3ce320-kube-api-access-jn9c7\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.505947 4559 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.505957 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.505965 4559 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.505993 4559 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b14d2018-932c-47b6-bba9-232dff3ce320-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.506002 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b14d2018-932c-47b6-bba9-232dff3ce320-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.506009 4559 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.506018 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0534e805-e3fd-4193-80ed-3c9fa69a041a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.506027 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45ls6\" (UniqueName: \"kubernetes.io/projected/0534e805-e3fd-4193-80ed-3c9fa69a041a-kube-api-access-45ls6\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.623870 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-k6zjg" event={"ID":"b14d2018-932c-47b6-bba9-232dff3ce320","Type":"ContainerDied","Data":"57f3f54f0ed9b335c62666e41423bd0c550cd93e89fb3684a1109a89c6cfd699"} Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.623910 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57f3f54f0ed9b335c62666e41423bd0c550cd93e89fb3684a1109a89c6cfd699" Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.623968 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-k6zjg" Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.626396 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hz99c" Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.629980 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hz99c" event={"ID":"0534e805-e3fd-4193-80ed-3c9fa69a041a","Type":"ContainerDied","Data":"e686b770243e7c35f771f1e539998d20e07bfbb18e5fb71c6af34d14c54cf1cc"} Nov 23 06:59:48 crc kubenswrapper[4559]: I1123 06:59:48.630030 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e686b770243e7c35f771f1e539998d20e07bfbb18e5fb71c6af34d14c54cf1cc" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.304278 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.305753 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.403725 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.469967 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.479259 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.525741 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-ovsdbserver-sb\") pod \"58898cc0-18e6-41c6-9ce2-149d081ff29b\" (UID: \"58898cc0-18e6-41c6-9ce2-149d081ff29b\") " Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.525851 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-ovsdbserver-nb\") pod \"58898cc0-18e6-41c6-9ce2-149d081ff29b\" (UID: \"58898cc0-18e6-41c6-9ce2-149d081ff29b\") " Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.525869 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-dns-swift-storage-0\") pod \"58898cc0-18e6-41c6-9ce2-149d081ff29b\" (UID: \"58898cc0-18e6-41c6-9ce2-149d081ff29b\") " Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.525898 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-config\") pod \"58898cc0-18e6-41c6-9ce2-149d081ff29b\" (UID: \"58898cc0-18e6-41c6-9ce2-149d081ff29b\") " Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.525933 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-dns-svc\") pod \"58898cc0-18e6-41c6-9ce2-149d081ff29b\" (UID: \"58898cc0-18e6-41c6-9ce2-149d081ff29b\") " Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.525958 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m77lh\" (UniqueName: \"kubernetes.io/projected/58898cc0-18e6-41c6-9ce2-149d081ff29b-kube-api-access-m77lh\") pod \"58898cc0-18e6-41c6-9ce2-149d081ff29b\" (UID: \"58898cc0-18e6-41c6-9ce2-149d081ff29b\") " Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.550132 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58898cc0-18e6-41c6-9ce2-149d081ff29b-kube-api-access-m77lh" (OuterVolumeSpecName: "kube-api-access-m77lh") pod "58898cc0-18e6-41c6-9ce2-149d081ff29b" (UID: "58898cc0-18e6-41c6-9ce2-149d081ff29b"). InnerVolumeSpecName "kube-api-access-m77lh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.587042 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-d4dd9b9cc-ndjpd"] Nov 23 06:59:49 crc kubenswrapper[4559]: E1123 06:59:49.587488 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58898cc0-18e6-41c6-9ce2-149d081ff29b" containerName="init" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.587507 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="58898cc0-18e6-41c6-9ce2-149d081ff29b" containerName="init" Nov 23 06:59:49 crc kubenswrapper[4559]: E1123 06:59:49.587527 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b14d2018-932c-47b6-bba9-232dff3ce320" containerName="barbican-db-sync" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.587534 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="b14d2018-932c-47b6-bba9-232dff3ce320" containerName="barbican-db-sync" Nov 23 06:59:49 crc kubenswrapper[4559]: E1123 06:59:49.587580 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0534e805-e3fd-4193-80ed-3c9fa69a041a" containerName="keystone-bootstrap" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.587586 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="0534e805-e3fd-4193-80ed-3c9fa69a041a" containerName="keystone-bootstrap" Nov 23 06:59:49 crc kubenswrapper[4559]: E1123 06:59:49.587601 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58898cc0-18e6-41c6-9ce2-149d081ff29b" containerName="dnsmasq-dns" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.587608 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="58898cc0-18e6-41c6-9ce2-149d081ff29b" containerName="dnsmasq-dns" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.587785 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="b14d2018-932c-47b6-bba9-232dff3ce320" containerName="barbican-db-sync" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.587800 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="58898cc0-18e6-41c6-9ce2-149d081ff29b" containerName="dnsmasq-dns" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.587811 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="0534e805-e3fd-4193-80ed-3c9fa69a041a" containerName="keystone-bootstrap" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.600410 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-d4dd9b9cc-ndjpd" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.647765 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m77lh\" (UniqueName: \"kubernetes.io/projected/58898cc0-18e6-41c6-9ce2-149d081ff29b-kube-api-access-m77lh\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.649962 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-lztfj" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.650523 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.650816 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.650941 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-685df6659d-zwp5c"] Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.652909 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-685df6659d-zwp5c" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.665297 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.676044 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-d4dd9b9cc-ndjpd"] Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.682292 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-config" (OuterVolumeSpecName: "config") pod "58898cc0-18e6-41c6-9ce2-149d081ff29b" (UID: "58898cc0-18e6-41c6-9ce2-149d081ff29b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.685141 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "58898cc0-18e6-41c6-9ce2-149d081ff29b" (UID: "58898cc0-18e6-41c6-9ce2-149d081ff29b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.703155 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.703405 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d78ff46f5-kdzjr" event={"ID":"58898cc0-18e6-41c6-9ce2-149d081ff29b","Type":"ContainerDied","Data":"c9e7ab3cf0b100c155267ee1c988ad1cc79bc7412ea16b6da38e6ceb7c86a95f"} Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.703450 4559 scope.go:117] "RemoveContainer" containerID="93860435ee52af0972f94019211071574356bbc2034f1651ff6d61882b1e7bbe" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.718743 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-685df6659d-zwp5c"] Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.721063 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "58898cc0-18e6-41c6-9ce2-149d081ff29b" (UID: "58898cc0-18e6-41c6-9ce2-149d081ff29b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.726102 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "58898cc0-18e6-41c6-9ce2-149d081ff29b" (UID: "58898cc0-18e6-41c6-9ce2-149d081ff29b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.738096 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "58898cc0-18e6-41c6-9ce2-149d081ff29b" (UID: "58898cc0-18e6-41c6-9ce2-149d081ff29b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.740866 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-74ff86f86c-gcbcl"] Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.742018 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-74ff86f86c-gcbcl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.744777 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xmhpd" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.744833 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.745500 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.746007 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.746206 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.746403 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.752704 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb7ba735-691f-4eb5-ac0c-c3f976e8e604-combined-ca-bundle\") pod \"barbican-keystone-listener-685df6659d-zwp5c\" (UID: \"cb7ba735-691f-4eb5-ac0c-c3f976e8e604\") " pod="openstack/barbican-keystone-listener-685df6659d-zwp5c" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.752788 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5b2ad32-159b-4369-951d-02d3ae8581eb-logs\") pod \"barbican-worker-d4dd9b9cc-ndjpd\" (UID: \"f5b2ad32-159b-4369-951d-02d3ae8581eb\") " pod="openstack/barbican-worker-d4dd9b9cc-ndjpd" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.752893 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5b2ad32-159b-4369-951d-02d3ae8581eb-combined-ca-bundle\") pod \"barbican-worker-d4dd9b9cc-ndjpd\" (UID: \"f5b2ad32-159b-4369-951d-02d3ae8581eb\") " pod="openstack/barbican-worker-d4dd9b9cc-ndjpd" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.752941 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb7ba735-691f-4eb5-ac0c-c3f976e8e604-config-data\") pod \"barbican-keystone-listener-685df6659d-zwp5c\" (UID: \"cb7ba735-691f-4eb5-ac0c-c3f976e8e604\") " pod="openstack/barbican-keystone-listener-685df6659d-zwp5c" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.752980 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb7ba735-691f-4eb5-ac0c-c3f976e8e604-logs\") pod \"barbican-keystone-listener-685df6659d-zwp5c\" (UID: \"cb7ba735-691f-4eb5-ac0c-c3f976e8e604\") " pod="openstack/barbican-keystone-listener-685df6659d-zwp5c" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.753046 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hznbg\" (UniqueName: \"kubernetes.io/projected/f5b2ad32-159b-4369-951d-02d3ae8581eb-kube-api-access-hznbg\") pod \"barbican-worker-d4dd9b9cc-ndjpd\" (UID: \"f5b2ad32-159b-4369-951d-02d3ae8581eb\") " pod="openstack/barbican-worker-d4dd9b9cc-ndjpd" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.753074 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5qfh\" (UniqueName: \"kubernetes.io/projected/cb7ba735-691f-4eb5-ac0c-c3f976e8e604-kube-api-access-x5qfh\") pod \"barbican-keystone-listener-685df6659d-zwp5c\" (UID: \"cb7ba735-691f-4eb5-ac0c-c3f976e8e604\") " pod="openstack/barbican-keystone-listener-685df6659d-zwp5c" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.753136 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5b2ad32-159b-4369-951d-02d3ae8581eb-config-data\") pod \"barbican-worker-d4dd9b9cc-ndjpd\" (UID: \"f5b2ad32-159b-4369-951d-02d3ae8581eb\") " pod="openstack/barbican-worker-d4dd9b9cc-ndjpd" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.753155 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb7ba735-691f-4eb5-ac0c-c3f976e8e604-config-data-custom\") pod \"barbican-keystone-listener-685df6659d-zwp5c\" (UID: \"cb7ba735-691f-4eb5-ac0c-c3f976e8e604\") " pod="openstack/barbican-keystone-listener-685df6659d-zwp5c" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.753195 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5b2ad32-159b-4369-951d-02d3ae8581eb-config-data-custom\") pod \"barbican-worker-d4dd9b9cc-ndjpd\" (UID: \"f5b2ad32-159b-4369-951d-02d3ae8581eb\") " pod="openstack/barbican-worker-d4dd9b9cc-ndjpd" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.753251 4559 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.753260 4559 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.753268 4559 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.753277 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.753286 4559 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58898cc0-18e6-41c6-9ce2-149d081ff29b-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.761840 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-74ff86f86c-gcbcl"] Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.803489 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-597c64895-wx4bl"] Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.805335 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-597c64895-wx4bl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.812250 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-597c64895-wx4bl"] Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.838041 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-798c5546f8-28xxs"] Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.839182 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-798c5546f8-28xxs" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.841205 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.852669 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-798c5546f8-28xxs"] Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.854727 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb7ba735-691f-4eb5-ac0c-c3f976e8e604-config-data\") pod \"barbican-keystone-listener-685df6659d-zwp5c\" (UID: \"cb7ba735-691f-4eb5-ac0c-c3f976e8e604\") " pod="openstack/barbican-keystone-listener-685df6659d-zwp5c" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.854764 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb7ba735-691f-4eb5-ac0c-c3f976e8e604-logs\") pod \"barbican-keystone-listener-685df6659d-zwp5c\" (UID: \"cb7ba735-691f-4eb5-ac0c-c3f976e8e604\") " pod="openstack/barbican-keystone-listener-685df6659d-zwp5c" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.854787 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dca8804f-80bf-4b44-99d0-7c2a9fea9166-scripts\") pod \"keystone-74ff86f86c-gcbcl\" (UID: \"dca8804f-80bf-4b44-99d0-7c2a9fea9166\") " pod="openstack/keystone-74ff86f86c-gcbcl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.854831 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dca8804f-80bf-4b44-99d0-7c2a9fea9166-credential-keys\") pod \"keystone-74ff86f86c-gcbcl\" (UID: \"dca8804f-80bf-4b44-99d0-7c2a9fea9166\") " pod="openstack/keystone-74ff86f86c-gcbcl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.854870 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hznbg\" (UniqueName: \"kubernetes.io/projected/f5b2ad32-159b-4369-951d-02d3ae8581eb-kube-api-access-hznbg\") pod \"barbican-worker-d4dd9b9cc-ndjpd\" (UID: \"f5b2ad32-159b-4369-951d-02d3ae8581eb\") " pod="openstack/barbican-worker-d4dd9b9cc-ndjpd" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.854891 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5qfh\" (UniqueName: \"kubernetes.io/projected/cb7ba735-691f-4eb5-ac0c-c3f976e8e604-kube-api-access-x5qfh\") pod \"barbican-keystone-listener-685df6659d-zwp5c\" (UID: \"cb7ba735-691f-4eb5-ac0c-c3f976e8e604\") " pod="openstack/barbican-keystone-listener-685df6659d-zwp5c" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.854945 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5b2ad32-159b-4369-951d-02d3ae8581eb-config-data\") pod \"barbican-worker-d4dd9b9cc-ndjpd\" (UID: \"f5b2ad32-159b-4369-951d-02d3ae8581eb\") " pod="openstack/barbican-worker-d4dd9b9cc-ndjpd" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.854960 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb7ba735-691f-4eb5-ac0c-c3f976e8e604-config-data-custom\") pod \"barbican-keystone-listener-685df6659d-zwp5c\" (UID: \"cb7ba735-691f-4eb5-ac0c-c3f976e8e604\") " pod="openstack/barbican-keystone-listener-685df6659d-zwp5c" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.854974 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dca8804f-80bf-4b44-99d0-7c2a9fea9166-config-data\") pod \"keystone-74ff86f86c-gcbcl\" (UID: \"dca8804f-80bf-4b44-99d0-7c2a9fea9166\") " pod="openstack/keystone-74ff86f86c-gcbcl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.855002 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dca8804f-80bf-4b44-99d0-7c2a9fea9166-public-tls-certs\") pod \"keystone-74ff86f86c-gcbcl\" (UID: \"dca8804f-80bf-4b44-99d0-7c2a9fea9166\") " pod="openstack/keystone-74ff86f86c-gcbcl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.855018 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lslfg\" (UniqueName: \"kubernetes.io/projected/dca8804f-80bf-4b44-99d0-7c2a9fea9166-kube-api-access-lslfg\") pod \"keystone-74ff86f86c-gcbcl\" (UID: \"dca8804f-80bf-4b44-99d0-7c2a9fea9166\") " pod="openstack/keystone-74ff86f86c-gcbcl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.855035 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5b2ad32-159b-4369-951d-02d3ae8581eb-config-data-custom\") pod \"barbican-worker-d4dd9b9cc-ndjpd\" (UID: \"f5b2ad32-159b-4369-951d-02d3ae8581eb\") " pod="openstack/barbican-worker-d4dd9b9cc-ndjpd" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.855051 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb7ba735-691f-4eb5-ac0c-c3f976e8e604-combined-ca-bundle\") pod \"barbican-keystone-listener-685df6659d-zwp5c\" (UID: \"cb7ba735-691f-4eb5-ac0c-c3f976e8e604\") " pod="openstack/barbican-keystone-listener-685df6659d-zwp5c" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.855070 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dca8804f-80bf-4b44-99d0-7c2a9fea9166-fernet-keys\") pod \"keystone-74ff86f86c-gcbcl\" (UID: \"dca8804f-80bf-4b44-99d0-7c2a9fea9166\") " pod="openstack/keystone-74ff86f86c-gcbcl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.855098 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5b2ad32-159b-4369-951d-02d3ae8581eb-logs\") pod \"barbican-worker-d4dd9b9cc-ndjpd\" (UID: \"f5b2ad32-159b-4369-951d-02d3ae8581eb\") " pod="openstack/barbican-worker-d4dd9b9cc-ndjpd" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.855114 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dca8804f-80bf-4b44-99d0-7c2a9fea9166-internal-tls-certs\") pod \"keystone-74ff86f86c-gcbcl\" (UID: \"dca8804f-80bf-4b44-99d0-7c2a9fea9166\") " pod="openstack/keystone-74ff86f86c-gcbcl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.855171 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dca8804f-80bf-4b44-99d0-7c2a9fea9166-combined-ca-bundle\") pod \"keystone-74ff86f86c-gcbcl\" (UID: \"dca8804f-80bf-4b44-99d0-7c2a9fea9166\") " pod="openstack/keystone-74ff86f86c-gcbcl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.855195 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5b2ad32-159b-4369-951d-02d3ae8581eb-combined-ca-bundle\") pod \"barbican-worker-d4dd9b9cc-ndjpd\" (UID: \"f5b2ad32-159b-4369-951d-02d3ae8581eb\") " pod="openstack/barbican-worker-d4dd9b9cc-ndjpd" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.859574 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5b2ad32-159b-4369-951d-02d3ae8581eb-logs\") pod \"barbican-worker-d4dd9b9cc-ndjpd\" (UID: \"f5b2ad32-159b-4369-951d-02d3ae8581eb\") " pod="openstack/barbican-worker-d4dd9b9cc-ndjpd" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.859973 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb7ba735-691f-4eb5-ac0c-c3f976e8e604-logs\") pod \"barbican-keystone-listener-685df6659d-zwp5c\" (UID: \"cb7ba735-691f-4eb5-ac0c-c3f976e8e604\") " pod="openstack/barbican-keystone-listener-685df6659d-zwp5c" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.860580 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5b2ad32-159b-4369-951d-02d3ae8581eb-combined-ca-bundle\") pod \"barbican-worker-d4dd9b9cc-ndjpd\" (UID: \"f5b2ad32-159b-4369-951d-02d3ae8581eb\") " pod="openstack/barbican-worker-d4dd9b9cc-ndjpd" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.863843 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5b2ad32-159b-4369-951d-02d3ae8581eb-config-data-custom\") pod \"barbican-worker-d4dd9b9cc-ndjpd\" (UID: \"f5b2ad32-159b-4369-951d-02d3ae8581eb\") " pod="openstack/barbican-worker-d4dd9b9cc-ndjpd" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.864670 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5b2ad32-159b-4369-951d-02d3ae8581eb-config-data\") pod \"barbican-worker-d4dd9b9cc-ndjpd\" (UID: \"f5b2ad32-159b-4369-951d-02d3ae8581eb\") " pod="openstack/barbican-worker-d4dd9b9cc-ndjpd" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.864752 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb7ba735-691f-4eb5-ac0c-c3f976e8e604-config-data\") pod \"barbican-keystone-listener-685df6659d-zwp5c\" (UID: \"cb7ba735-691f-4eb5-ac0c-c3f976e8e604\") " pod="openstack/barbican-keystone-listener-685df6659d-zwp5c" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.868577 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb7ba735-691f-4eb5-ac0c-c3f976e8e604-combined-ca-bundle\") pod \"barbican-keystone-listener-685df6659d-zwp5c\" (UID: \"cb7ba735-691f-4eb5-ac0c-c3f976e8e604\") " pod="openstack/barbican-keystone-listener-685df6659d-zwp5c" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.871797 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb7ba735-691f-4eb5-ac0c-c3f976e8e604-config-data-custom\") pod \"barbican-keystone-listener-685df6659d-zwp5c\" (UID: \"cb7ba735-691f-4eb5-ac0c-c3f976e8e604\") " pod="openstack/barbican-keystone-listener-685df6659d-zwp5c" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.873526 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hznbg\" (UniqueName: \"kubernetes.io/projected/f5b2ad32-159b-4369-951d-02d3ae8581eb-kube-api-access-hznbg\") pod \"barbican-worker-d4dd9b9cc-ndjpd\" (UID: \"f5b2ad32-159b-4369-951d-02d3ae8581eb\") " pod="openstack/barbican-worker-d4dd9b9cc-ndjpd" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.877671 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5qfh\" (UniqueName: \"kubernetes.io/projected/cb7ba735-691f-4eb5-ac0c-c3f976e8e604-kube-api-access-x5qfh\") pod \"barbican-keystone-listener-685df6659d-zwp5c\" (UID: \"cb7ba735-691f-4eb5-ac0c-c3f976e8e604\") " pod="openstack/barbican-keystone-listener-685df6659d-zwp5c" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.957351 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dklj7\" (UniqueName: \"kubernetes.io/projected/94879e46-753f-4135-a501-c9ff00800450-kube-api-access-dklj7\") pod \"dnsmasq-dns-597c64895-wx4bl\" (UID: \"94879e46-753f-4135-a501-c9ff00800450\") " pod="openstack/dnsmasq-dns-597c64895-wx4bl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.958377 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-config\") pod \"dnsmasq-dns-597c64895-wx4bl\" (UID: \"94879e46-753f-4135-a501-c9ff00800450\") " pod="openstack/dnsmasq-dns-597c64895-wx4bl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.958477 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-ovsdbserver-nb\") pod \"dnsmasq-dns-597c64895-wx4bl\" (UID: \"94879e46-753f-4135-a501-c9ff00800450\") " pod="openstack/dnsmasq-dns-597c64895-wx4bl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.958577 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69cec8df-ef13-4a2a-884b-78ba11c34b3e-config-data-custom\") pod \"barbican-api-798c5546f8-28xxs\" (UID: \"69cec8df-ef13-4a2a-884b-78ba11c34b3e\") " pod="openstack/barbican-api-798c5546f8-28xxs" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.958687 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-dns-svc\") pod \"dnsmasq-dns-597c64895-wx4bl\" (UID: \"94879e46-753f-4135-a501-c9ff00800450\") " pod="openstack/dnsmasq-dns-597c64895-wx4bl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.958786 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-ovsdbserver-sb\") pod \"dnsmasq-dns-597c64895-wx4bl\" (UID: \"94879e46-753f-4135-a501-c9ff00800450\") " pod="openstack/dnsmasq-dns-597c64895-wx4bl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.958852 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dca8804f-80bf-4b44-99d0-7c2a9fea9166-scripts\") pod \"keystone-74ff86f86c-gcbcl\" (UID: \"dca8804f-80bf-4b44-99d0-7c2a9fea9166\") " pod="openstack/keystone-74ff86f86c-gcbcl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.958956 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dca8804f-80bf-4b44-99d0-7c2a9fea9166-credential-keys\") pod \"keystone-74ff86f86c-gcbcl\" (UID: \"dca8804f-80bf-4b44-99d0-7c2a9fea9166\") " pod="openstack/keystone-74ff86f86c-gcbcl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.959125 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-dns-swift-storage-0\") pod \"dnsmasq-dns-597c64895-wx4bl\" (UID: \"94879e46-753f-4135-a501-c9ff00800450\") " pod="openstack/dnsmasq-dns-597c64895-wx4bl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.959185 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bttvb\" (UniqueName: \"kubernetes.io/projected/69cec8df-ef13-4a2a-884b-78ba11c34b3e-kube-api-access-bttvb\") pod \"barbican-api-798c5546f8-28xxs\" (UID: \"69cec8df-ef13-4a2a-884b-78ba11c34b3e\") " pod="openstack/barbican-api-798c5546f8-28xxs" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.959269 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dca8804f-80bf-4b44-99d0-7c2a9fea9166-config-data\") pod \"keystone-74ff86f86c-gcbcl\" (UID: \"dca8804f-80bf-4b44-99d0-7c2a9fea9166\") " pod="openstack/keystone-74ff86f86c-gcbcl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.959353 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dca8804f-80bf-4b44-99d0-7c2a9fea9166-public-tls-certs\") pod \"keystone-74ff86f86c-gcbcl\" (UID: \"dca8804f-80bf-4b44-99d0-7c2a9fea9166\") " pod="openstack/keystone-74ff86f86c-gcbcl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.959415 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69cec8df-ef13-4a2a-884b-78ba11c34b3e-logs\") pod \"barbican-api-798c5546f8-28xxs\" (UID: \"69cec8df-ef13-4a2a-884b-78ba11c34b3e\") " pod="openstack/barbican-api-798c5546f8-28xxs" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.959479 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lslfg\" (UniqueName: \"kubernetes.io/projected/dca8804f-80bf-4b44-99d0-7c2a9fea9166-kube-api-access-lslfg\") pod \"keystone-74ff86f86c-gcbcl\" (UID: \"dca8804f-80bf-4b44-99d0-7c2a9fea9166\") " pod="openstack/keystone-74ff86f86c-gcbcl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.959539 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69cec8df-ef13-4a2a-884b-78ba11c34b3e-config-data\") pod \"barbican-api-798c5546f8-28xxs\" (UID: \"69cec8df-ef13-4a2a-884b-78ba11c34b3e\") " pod="openstack/barbican-api-798c5546f8-28xxs" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.959622 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dca8804f-80bf-4b44-99d0-7c2a9fea9166-fernet-keys\") pod \"keystone-74ff86f86c-gcbcl\" (UID: \"dca8804f-80bf-4b44-99d0-7c2a9fea9166\") " pod="openstack/keystone-74ff86f86c-gcbcl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.959725 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69cec8df-ef13-4a2a-884b-78ba11c34b3e-combined-ca-bundle\") pod \"barbican-api-798c5546f8-28xxs\" (UID: \"69cec8df-ef13-4a2a-884b-78ba11c34b3e\") " pod="openstack/barbican-api-798c5546f8-28xxs" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.959797 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dca8804f-80bf-4b44-99d0-7c2a9fea9166-internal-tls-certs\") pod \"keystone-74ff86f86c-gcbcl\" (UID: \"dca8804f-80bf-4b44-99d0-7c2a9fea9166\") " pod="openstack/keystone-74ff86f86c-gcbcl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.959913 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dca8804f-80bf-4b44-99d0-7c2a9fea9166-combined-ca-bundle\") pod \"keystone-74ff86f86c-gcbcl\" (UID: \"dca8804f-80bf-4b44-99d0-7c2a9fea9166\") " pod="openstack/keystone-74ff86f86c-gcbcl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.963853 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dca8804f-80bf-4b44-99d0-7c2a9fea9166-scripts\") pod \"keystone-74ff86f86c-gcbcl\" (UID: \"dca8804f-80bf-4b44-99d0-7c2a9fea9166\") " pod="openstack/keystone-74ff86f86c-gcbcl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.967527 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dca8804f-80bf-4b44-99d0-7c2a9fea9166-combined-ca-bundle\") pod \"keystone-74ff86f86c-gcbcl\" (UID: \"dca8804f-80bf-4b44-99d0-7c2a9fea9166\") " pod="openstack/keystone-74ff86f86c-gcbcl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.970052 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dca8804f-80bf-4b44-99d0-7c2a9fea9166-credential-keys\") pod \"keystone-74ff86f86c-gcbcl\" (UID: \"dca8804f-80bf-4b44-99d0-7c2a9fea9166\") " pod="openstack/keystone-74ff86f86c-gcbcl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.970063 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dca8804f-80bf-4b44-99d0-7c2a9fea9166-fernet-keys\") pod \"keystone-74ff86f86c-gcbcl\" (UID: \"dca8804f-80bf-4b44-99d0-7c2a9fea9166\") " pod="openstack/keystone-74ff86f86c-gcbcl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.970166 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dca8804f-80bf-4b44-99d0-7c2a9fea9166-internal-tls-certs\") pod \"keystone-74ff86f86c-gcbcl\" (UID: \"dca8804f-80bf-4b44-99d0-7c2a9fea9166\") " pod="openstack/keystone-74ff86f86c-gcbcl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.970425 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dca8804f-80bf-4b44-99d0-7c2a9fea9166-config-data\") pod \"keystone-74ff86f86c-gcbcl\" (UID: \"dca8804f-80bf-4b44-99d0-7c2a9fea9166\") " pod="openstack/keystone-74ff86f86c-gcbcl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.974032 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dca8804f-80bf-4b44-99d0-7c2a9fea9166-public-tls-certs\") pod \"keystone-74ff86f86c-gcbcl\" (UID: \"dca8804f-80bf-4b44-99d0-7c2a9fea9166\") " pod="openstack/keystone-74ff86f86c-gcbcl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.977091 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lslfg\" (UniqueName: \"kubernetes.io/projected/dca8804f-80bf-4b44-99d0-7c2a9fea9166-kube-api-access-lslfg\") pod \"keystone-74ff86f86c-gcbcl\" (UID: \"dca8804f-80bf-4b44-99d0-7c2a9fea9166\") " pod="openstack/keystone-74ff86f86c-gcbcl" Nov 23 06:59:49 crc kubenswrapper[4559]: I1123 06:59:49.987975 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-d4dd9b9cc-ndjpd" Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.015230 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-685df6659d-zwp5c" Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.036998 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d78ff46f5-kdzjr"] Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.041622 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d78ff46f5-kdzjr"] Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.057884 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-74ff86f86c-gcbcl" Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.061360 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-ovsdbserver-sb\") pod \"dnsmasq-dns-597c64895-wx4bl\" (UID: \"94879e46-753f-4135-a501-c9ff00800450\") " pod="openstack/dnsmasq-dns-597c64895-wx4bl" Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.061444 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-dns-swift-storage-0\") pod \"dnsmasq-dns-597c64895-wx4bl\" (UID: \"94879e46-753f-4135-a501-c9ff00800450\") " pod="openstack/dnsmasq-dns-597c64895-wx4bl" Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.061472 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bttvb\" (UniqueName: \"kubernetes.io/projected/69cec8df-ef13-4a2a-884b-78ba11c34b3e-kube-api-access-bttvb\") pod \"barbican-api-798c5546f8-28xxs\" (UID: \"69cec8df-ef13-4a2a-884b-78ba11c34b3e\") " pod="openstack/barbican-api-798c5546f8-28xxs" Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.061509 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69cec8df-ef13-4a2a-884b-78ba11c34b3e-logs\") pod \"barbican-api-798c5546f8-28xxs\" (UID: \"69cec8df-ef13-4a2a-884b-78ba11c34b3e\") " pod="openstack/barbican-api-798c5546f8-28xxs" Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.061528 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69cec8df-ef13-4a2a-884b-78ba11c34b3e-config-data\") pod \"barbican-api-798c5546f8-28xxs\" (UID: \"69cec8df-ef13-4a2a-884b-78ba11c34b3e\") " pod="openstack/barbican-api-798c5546f8-28xxs" Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.061560 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69cec8df-ef13-4a2a-884b-78ba11c34b3e-combined-ca-bundle\") pod \"barbican-api-798c5546f8-28xxs\" (UID: \"69cec8df-ef13-4a2a-884b-78ba11c34b3e\") " pod="openstack/barbican-api-798c5546f8-28xxs" Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.061608 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dklj7\" (UniqueName: \"kubernetes.io/projected/94879e46-753f-4135-a501-c9ff00800450-kube-api-access-dklj7\") pod \"dnsmasq-dns-597c64895-wx4bl\" (UID: \"94879e46-753f-4135-a501-c9ff00800450\") " pod="openstack/dnsmasq-dns-597c64895-wx4bl" Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.061632 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-config\") pod \"dnsmasq-dns-597c64895-wx4bl\" (UID: \"94879e46-753f-4135-a501-c9ff00800450\") " pod="openstack/dnsmasq-dns-597c64895-wx4bl" Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.061667 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-ovsdbserver-nb\") pod \"dnsmasq-dns-597c64895-wx4bl\" (UID: \"94879e46-753f-4135-a501-c9ff00800450\") " pod="openstack/dnsmasq-dns-597c64895-wx4bl" Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.061691 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69cec8df-ef13-4a2a-884b-78ba11c34b3e-config-data-custom\") pod \"barbican-api-798c5546f8-28xxs\" (UID: \"69cec8df-ef13-4a2a-884b-78ba11c34b3e\") " pod="openstack/barbican-api-798c5546f8-28xxs" Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.061711 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-dns-svc\") pod \"dnsmasq-dns-597c64895-wx4bl\" (UID: \"94879e46-753f-4135-a501-c9ff00800450\") " pod="openstack/dnsmasq-dns-597c64895-wx4bl" Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.062429 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-dns-svc\") pod \"dnsmasq-dns-597c64895-wx4bl\" (UID: \"94879e46-753f-4135-a501-c9ff00800450\") " pod="openstack/dnsmasq-dns-597c64895-wx4bl" Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.063054 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-ovsdbserver-sb\") pod \"dnsmasq-dns-597c64895-wx4bl\" (UID: \"94879e46-753f-4135-a501-c9ff00800450\") " pod="openstack/dnsmasq-dns-597c64895-wx4bl" Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.066010 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69cec8df-ef13-4a2a-884b-78ba11c34b3e-logs\") pod \"barbican-api-798c5546f8-28xxs\" (UID: \"69cec8df-ef13-4a2a-884b-78ba11c34b3e\") " pod="openstack/barbican-api-798c5546f8-28xxs" Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.066378 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-dns-swift-storage-0\") pod \"dnsmasq-dns-597c64895-wx4bl\" (UID: \"94879e46-753f-4135-a501-c9ff00800450\") " pod="openstack/dnsmasq-dns-597c64895-wx4bl" Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.066988 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-config\") pod \"dnsmasq-dns-597c64895-wx4bl\" (UID: \"94879e46-753f-4135-a501-c9ff00800450\") " pod="openstack/dnsmasq-dns-597c64895-wx4bl" Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.068313 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-ovsdbserver-nb\") pod \"dnsmasq-dns-597c64895-wx4bl\" (UID: \"94879e46-753f-4135-a501-c9ff00800450\") " pod="openstack/dnsmasq-dns-597c64895-wx4bl" Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.068456 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69cec8df-ef13-4a2a-884b-78ba11c34b3e-combined-ca-bundle\") pod \"barbican-api-798c5546f8-28xxs\" (UID: \"69cec8df-ef13-4a2a-884b-78ba11c34b3e\") " pod="openstack/barbican-api-798c5546f8-28xxs" Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.069904 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69cec8df-ef13-4a2a-884b-78ba11c34b3e-config-data\") pod \"barbican-api-798c5546f8-28xxs\" (UID: \"69cec8df-ef13-4a2a-884b-78ba11c34b3e\") " pod="openstack/barbican-api-798c5546f8-28xxs" Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.078716 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69cec8df-ef13-4a2a-884b-78ba11c34b3e-config-data-custom\") pod \"barbican-api-798c5546f8-28xxs\" (UID: \"69cec8df-ef13-4a2a-884b-78ba11c34b3e\") " pod="openstack/barbican-api-798c5546f8-28xxs" Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.079145 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bttvb\" (UniqueName: \"kubernetes.io/projected/69cec8df-ef13-4a2a-884b-78ba11c34b3e-kube-api-access-bttvb\") pod \"barbican-api-798c5546f8-28xxs\" (UID: \"69cec8df-ef13-4a2a-884b-78ba11c34b3e\") " pod="openstack/barbican-api-798c5546f8-28xxs" Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.080936 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dklj7\" (UniqueName: \"kubernetes.io/projected/94879e46-753f-4135-a501-c9ff00800450-kube-api-access-dklj7\") pod \"dnsmasq-dns-597c64895-wx4bl\" (UID: \"94879e46-753f-4135-a501-c9ff00800450\") " pod="openstack/dnsmasq-dns-597c64895-wx4bl" Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.138556 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-597c64895-wx4bl" Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.150860 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-798c5546f8-28xxs" Nov 23 06:59:50 crc kubenswrapper[4559]: I1123 06:59:50.284683 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58898cc0-18e6-41c6-9ce2-149d081ff29b" path="/var/lib/kubelet/pods/58898cc0-18e6-41c6-9ce2-149d081ff29b/volumes" Nov 23 06:59:51 crc kubenswrapper[4559]: I1123 06:59:51.154730 4559 scope.go:117] "RemoveContainer" containerID="75879545707b8bec91ac310eb56a716157422ee90fb701af77ed451048b5e4c9" Nov 23 06:59:51 crc kubenswrapper[4559]: I1123 06:59:51.554561 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7667f44776-mszxt"] Nov 23 06:59:51 crc kubenswrapper[4559]: W1123 06:59:51.560449 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1df68d96_dfae_42c9_8a11_9906486ea1b9.slice/crio-3719132a459aace6c82d04fc12df1a231e38ba0a94f5a5c87a04171e1887a2c8 WatchSource:0}: Error finding container 3719132a459aace6c82d04fc12df1a231e38ba0a94f5a5c87a04171e1887a2c8: Status 404 returned error can't find the container with id 3719132a459aace6c82d04fc12df1a231e38ba0a94f5a5c87a04171e1887a2c8 Nov 23 06:59:51 crc kubenswrapper[4559]: I1123 06:59:51.732108 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-685df6659d-zwp5c"] Nov 23 06:59:51 crc kubenswrapper[4559]: I1123 06:59:51.747015 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-74ff86f86c-gcbcl"] Nov 23 06:59:51 crc kubenswrapper[4559]: I1123 06:59:51.753378 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7667f44776-mszxt" event={"ID":"1df68d96-dfae-42c9-8a11-9906486ea1b9","Type":"ContainerStarted","Data":"3719132a459aace6c82d04fc12df1a231e38ba0a94f5a5c87a04171e1887a2c8"} Nov 23 06:59:51 crc kubenswrapper[4559]: I1123 06:59:51.754787 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-55bb64d698-6lpkd"] Nov 23 06:59:51 crc kubenswrapper[4559]: I1123 06:59:51.756222 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-55bb64d698-6lpkd" Nov 23 06:59:51 crc kubenswrapper[4559]: I1123 06:59:51.761723 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 23 06:59:51 crc kubenswrapper[4559]: I1123 06:59:51.761989 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 23 06:59:51 crc kubenswrapper[4559]: I1123 06:59:51.766118 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4163bfe8-0de7-4490-b825-55d0f0343787","Type":"ContainerStarted","Data":"5fb7c20c88f83d97306e5172d980cc901c730345621912965ed251efeb54307a"} Nov 23 06:59:51 crc kubenswrapper[4559]: I1123 06:59:51.768840 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-55bb64d698-6lpkd"] Nov 23 06:59:51 crc kubenswrapper[4559]: I1123 06:59:51.868630 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-d4dd9b9cc-ndjpd"] Nov 23 06:59:51 crc kubenswrapper[4559]: W1123 06:59:51.875767 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5b2ad32_159b_4369_951d_02d3ae8581eb.slice/crio-5aa3279aa9172b377992aafa5ed94b835ebcf315f99f57bb602f069ef5bf599f WatchSource:0}: Error finding container 5aa3279aa9172b377992aafa5ed94b835ebcf315f99f57bb602f069ef5bf599f: Status 404 returned error can't find the container with id 5aa3279aa9172b377992aafa5ed94b835ebcf315f99f57bb602f069ef5bf599f Nov 23 06:59:51 crc kubenswrapper[4559]: W1123 06:59:51.878057 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69cec8df_ef13_4a2a_884b_78ba11c34b3e.slice/crio-0a93f15632ff4ff14ff7816dae4e4f21250c232e2843dc579ae2530b9144b12b WatchSource:0}: Error finding container 0a93f15632ff4ff14ff7816dae4e4f21250c232e2843dc579ae2530b9144b12b: Status 404 returned error can't find the container with id 0a93f15632ff4ff14ff7816dae4e4f21250c232e2843dc579ae2530b9144b12b Nov 23 06:59:51 crc kubenswrapper[4559]: I1123 06:59:51.880218 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-798c5546f8-28xxs"] Nov 23 06:59:51 crc kubenswrapper[4559]: I1123 06:59:51.897347 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9264f09-464d-404a-88a1-7f86ecb2ccb8-logs\") pod \"barbican-api-55bb64d698-6lpkd\" (UID: \"f9264f09-464d-404a-88a1-7f86ecb2ccb8\") " pod="openstack/barbican-api-55bb64d698-6lpkd" Nov 23 06:59:51 crc kubenswrapper[4559]: I1123 06:59:51.897384 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9264f09-464d-404a-88a1-7f86ecb2ccb8-config-data\") pod \"barbican-api-55bb64d698-6lpkd\" (UID: \"f9264f09-464d-404a-88a1-7f86ecb2ccb8\") " pod="openstack/barbican-api-55bb64d698-6lpkd" Nov 23 06:59:51 crc kubenswrapper[4559]: I1123 06:59:51.897443 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f9264f09-464d-404a-88a1-7f86ecb2ccb8-config-data-custom\") pod \"barbican-api-55bb64d698-6lpkd\" (UID: \"f9264f09-464d-404a-88a1-7f86ecb2ccb8\") " pod="openstack/barbican-api-55bb64d698-6lpkd" Nov 23 06:59:51 crc kubenswrapper[4559]: I1123 06:59:51.897528 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9264f09-464d-404a-88a1-7f86ecb2ccb8-public-tls-certs\") pod \"barbican-api-55bb64d698-6lpkd\" (UID: \"f9264f09-464d-404a-88a1-7f86ecb2ccb8\") " pod="openstack/barbican-api-55bb64d698-6lpkd" Nov 23 06:59:51 crc kubenswrapper[4559]: I1123 06:59:51.897583 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9264f09-464d-404a-88a1-7f86ecb2ccb8-combined-ca-bundle\") pod \"barbican-api-55bb64d698-6lpkd\" (UID: \"f9264f09-464d-404a-88a1-7f86ecb2ccb8\") " pod="openstack/barbican-api-55bb64d698-6lpkd" Nov 23 06:59:51 crc kubenswrapper[4559]: I1123 06:59:51.897706 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgws2\" (UniqueName: \"kubernetes.io/projected/f9264f09-464d-404a-88a1-7f86ecb2ccb8-kube-api-access-tgws2\") pod \"barbican-api-55bb64d698-6lpkd\" (UID: \"f9264f09-464d-404a-88a1-7f86ecb2ccb8\") " pod="openstack/barbican-api-55bb64d698-6lpkd" Nov 23 06:59:51 crc kubenswrapper[4559]: I1123 06:59:51.897800 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9264f09-464d-404a-88a1-7f86ecb2ccb8-internal-tls-certs\") pod \"barbican-api-55bb64d698-6lpkd\" (UID: \"f9264f09-464d-404a-88a1-7f86ecb2ccb8\") " pod="openstack/barbican-api-55bb64d698-6lpkd" Nov 23 06:59:51 crc kubenswrapper[4559]: I1123 06:59:51.999309 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9264f09-464d-404a-88a1-7f86ecb2ccb8-combined-ca-bundle\") pod \"barbican-api-55bb64d698-6lpkd\" (UID: \"f9264f09-464d-404a-88a1-7f86ecb2ccb8\") " pod="openstack/barbican-api-55bb64d698-6lpkd" Nov 23 06:59:51 crc kubenswrapper[4559]: I1123 06:59:51.999418 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgws2\" (UniqueName: \"kubernetes.io/projected/f9264f09-464d-404a-88a1-7f86ecb2ccb8-kube-api-access-tgws2\") pod \"barbican-api-55bb64d698-6lpkd\" (UID: \"f9264f09-464d-404a-88a1-7f86ecb2ccb8\") " pod="openstack/barbican-api-55bb64d698-6lpkd" Nov 23 06:59:51 crc kubenswrapper[4559]: I1123 06:59:51.999758 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9264f09-464d-404a-88a1-7f86ecb2ccb8-internal-tls-certs\") pod \"barbican-api-55bb64d698-6lpkd\" (UID: \"f9264f09-464d-404a-88a1-7f86ecb2ccb8\") " pod="openstack/barbican-api-55bb64d698-6lpkd" Nov 23 06:59:51 crc kubenswrapper[4559]: I1123 06:59:51.999795 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9264f09-464d-404a-88a1-7f86ecb2ccb8-logs\") pod \"barbican-api-55bb64d698-6lpkd\" (UID: \"f9264f09-464d-404a-88a1-7f86ecb2ccb8\") " pod="openstack/barbican-api-55bb64d698-6lpkd" Nov 23 06:59:51 crc kubenswrapper[4559]: I1123 06:59:51.999815 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9264f09-464d-404a-88a1-7f86ecb2ccb8-config-data\") pod \"barbican-api-55bb64d698-6lpkd\" (UID: \"f9264f09-464d-404a-88a1-7f86ecb2ccb8\") " pod="openstack/barbican-api-55bb64d698-6lpkd" Nov 23 06:59:51 crc kubenswrapper[4559]: I1123 06:59:51.999853 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f9264f09-464d-404a-88a1-7f86ecb2ccb8-config-data-custom\") pod \"barbican-api-55bb64d698-6lpkd\" (UID: \"f9264f09-464d-404a-88a1-7f86ecb2ccb8\") " pod="openstack/barbican-api-55bb64d698-6lpkd" Nov 23 06:59:51 crc kubenswrapper[4559]: I1123 06:59:51.999886 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9264f09-464d-404a-88a1-7f86ecb2ccb8-public-tls-certs\") pod \"barbican-api-55bb64d698-6lpkd\" (UID: \"f9264f09-464d-404a-88a1-7f86ecb2ccb8\") " pod="openstack/barbican-api-55bb64d698-6lpkd" Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.003361 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-597c64895-wx4bl"] Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.005437 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f9264f09-464d-404a-88a1-7f86ecb2ccb8-config-data-custom\") pod \"barbican-api-55bb64d698-6lpkd\" (UID: \"f9264f09-464d-404a-88a1-7f86ecb2ccb8\") " pod="openstack/barbican-api-55bb64d698-6lpkd" Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.006900 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9264f09-464d-404a-88a1-7f86ecb2ccb8-logs\") pod \"barbican-api-55bb64d698-6lpkd\" (UID: \"f9264f09-464d-404a-88a1-7f86ecb2ccb8\") " pod="openstack/barbican-api-55bb64d698-6lpkd" Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.010877 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9264f09-464d-404a-88a1-7f86ecb2ccb8-public-tls-certs\") pod \"barbican-api-55bb64d698-6lpkd\" (UID: \"f9264f09-464d-404a-88a1-7f86ecb2ccb8\") " pod="openstack/barbican-api-55bb64d698-6lpkd" Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.011732 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9264f09-464d-404a-88a1-7f86ecb2ccb8-config-data\") pod \"barbican-api-55bb64d698-6lpkd\" (UID: \"f9264f09-464d-404a-88a1-7f86ecb2ccb8\") " pod="openstack/barbican-api-55bb64d698-6lpkd" Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.013202 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9264f09-464d-404a-88a1-7f86ecb2ccb8-combined-ca-bundle\") pod \"barbican-api-55bb64d698-6lpkd\" (UID: \"f9264f09-464d-404a-88a1-7f86ecb2ccb8\") " pod="openstack/barbican-api-55bb64d698-6lpkd" Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.013412 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9264f09-464d-404a-88a1-7f86ecb2ccb8-internal-tls-certs\") pod \"barbican-api-55bb64d698-6lpkd\" (UID: \"f9264f09-464d-404a-88a1-7f86ecb2ccb8\") " pod="openstack/barbican-api-55bb64d698-6lpkd" Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.014934 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgws2\" (UniqueName: \"kubernetes.io/projected/f9264f09-464d-404a-88a1-7f86ecb2ccb8-kube-api-access-tgws2\") pod \"barbican-api-55bb64d698-6lpkd\" (UID: \"f9264f09-464d-404a-88a1-7f86ecb2ccb8\") " pod="openstack/barbican-api-55bb64d698-6lpkd" Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.071551 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-55bb64d698-6lpkd" Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.498133 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-55bb64d698-6lpkd"] Nov 23 06:59:52 crc kubenswrapper[4559]: W1123 06:59:52.508566 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9264f09_464d_404a_88a1_7f86ecb2ccb8.slice/crio-19eb56301b00e0949cf81932b094fdc333c4059fcc1f2041287eb11cec939d88 WatchSource:0}: Error finding container 19eb56301b00e0949cf81932b094fdc333c4059fcc1f2041287eb11cec939d88: Status 404 returned error can't find the container with id 19eb56301b00e0949cf81932b094fdc333c4059fcc1f2041287eb11cec939d88 Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.778541 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-d4dd9b9cc-ndjpd" event={"ID":"f5b2ad32-159b-4369-951d-02d3ae8581eb","Type":"ContainerStarted","Data":"5aa3279aa9172b377992aafa5ed94b835ebcf315f99f57bb602f069ef5bf599f"} Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.779973 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-685df6659d-zwp5c" event={"ID":"cb7ba735-691f-4eb5-ac0c-c3f976e8e604","Type":"ContainerStarted","Data":"6f8d85d745e6ed2ed9eb78b65d03e4c2d47394a46c43b4619f79c55e54998fae"} Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.781784 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7667f44776-mszxt" event={"ID":"1df68d96-dfae-42c9-8a11-9906486ea1b9","Type":"ContainerStarted","Data":"4f6133bad59ef7e0b733c82f3375e48b64a8a2647def2a2bb439e4c547806970"} Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.781810 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7667f44776-mszxt" event={"ID":"1df68d96-dfae-42c9-8a11-9906486ea1b9","Type":"ContainerStarted","Data":"461d0c0ed9472dacbd1f6384ee278cc471f5dd01b9a3018a404caa20196d5bf1"} Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.781893 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7667f44776-mszxt" Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.781975 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7667f44776-mszxt" Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.783946 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-74ff86f86c-gcbcl" event={"ID":"dca8804f-80bf-4b44-99d0-7c2a9fea9166","Type":"ContainerStarted","Data":"f57840fd891f79510247a841541733a555fa8c7c3c684bc0108cbb55b93b9251"} Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.783977 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-74ff86f86c-gcbcl" event={"ID":"dca8804f-80bf-4b44-99d0-7c2a9fea9166","Type":"ContainerStarted","Data":"34bb04b90649dfe856793c6833a7e1ed4cb8515aa1f02ba02b32c8098fadc261"} Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.784092 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-74ff86f86c-gcbcl" Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.785355 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-55bb64d698-6lpkd" event={"ID":"f9264f09-464d-404a-88a1-7f86ecb2ccb8","Type":"ContainerStarted","Data":"19eb56301b00e0949cf81932b094fdc333c4059fcc1f2041287eb11cec939d88"} Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.787219 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-798c5546f8-28xxs" event={"ID":"69cec8df-ef13-4a2a-884b-78ba11c34b3e","Type":"ContainerStarted","Data":"1c44ab5eb8f279614cdf3f57ebe976f6d3d88559796aa8cd24bd1e1462530a00"} Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.787256 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-798c5546f8-28xxs" event={"ID":"69cec8df-ef13-4a2a-884b-78ba11c34b3e","Type":"ContainerStarted","Data":"b9714d8b8006b6c8af8b3d0c7c2cf22f1e859448a454ee2e74ac2bf1b2c6fcee"} Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.787270 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-798c5546f8-28xxs" event={"ID":"69cec8df-ef13-4a2a-884b-78ba11c34b3e","Type":"ContainerStarted","Data":"0a93f15632ff4ff14ff7816dae4e4f21250c232e2843dc579ae2530b9144b12b"} Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.787315 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-798c5546f8-28xxs" Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.787365 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-798c5546f8-28xxs" Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.789041 4559 generic.go:334] "Generic (PLEG): container finished" podID="94879e46-753f-4135-a501-c9ff00800450" containerID="e505ab11f0d0748ac9c3d7c30dfcda70af23e0a4d8b39029599104affd2c85ef" exitCode=0 Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.789121 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-597c64895-wx4bl" event={"ID":"94879e46-753f-4135-a501-c9ff00800450","Type":"ContainerDied","Data":"e505ab11f0d0748ac9c3d7c30dfcda70af23e0a4d8b39029599104affd2c85ef"} Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.789190 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-597c64895-wx4bl" event={"ID":"94879e46-753f-4135-a501-c9ff00800450","Type":"ContainerStarted","Data":"a315c7cb9f283bb090318c6799ff819ff6be2ddb5350db39d4500ecb1aed3292"} Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.790701 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-bxpjp" event={"ID":"c8ae1721-4ded-4e7d-9091-88ac908d0554","Type":"ContainerStarted","Data":"178d1c0f29794efd9880d96375c09387602578935cfa41dca7f541e19feea9ea"} Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.803333 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-7667f44776-mszxt" podStartSLOduration=12.803311546 podStartE2EDuration="12.803311546s" podCreationTimestamp="2025-11-23 06:59:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:59:52.795793206 +0000 UTC m=+894.817778820" watchObservedRunningTime="2025-11-23 06:59:52.803311546 +0000 UTC m=+894.825297160" Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.830277 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-74ff86f86c-gcbcl" podStartSLOduration=3.83025567 podStartE2EDuration="3.83025567s" podCreationTimestamp="2025-11-23 06:59:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:59:52.823223185 +0000 UTC m=+894.845208798" watchObservedRunningTime="2025-11-23 06:59:52.83025567 +0000 UTC m=+894.852241304" Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.850547 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-bxpjp" podStartSLOduration=3.547428113 podStartE2EDuration="27.850531634s" podCreationTimestamp="2025-11-23 06:59:25 +0000 UTC" firstStartedPulling="2025-11-23 06:59:26.877515753 +0000 UTC m=+868.899501367" lastFinishedPulling="2025-11-23 06:59:51.180619274 +0000 UTC m=+893.202604888" observedRunningTime="2025-11-23 06:59:52.841177612 +0000 UTC m=+894.863163226" watchObservedRunningTime="2025-11-23 06:59:52.850531634 +0000 UTC m=+894.872517248" Nov 23 06:59:52 crc kubenswrapper[4559]: I1123 06:59:52.857423 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-798c5546f8-28xxs" podStartSLOduration=3.857417264 podStartE2EDuration="3.857417264s" podCreationTimestamp="2025-11-23 06:59:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:59:52.854594515 +0000 UTC m=+894.876580129" watchObservedRunningTime="2025-11-23 06:59:52.857417264 +0000 UTC m=+894.879402878" Nov 23 06:59:53 crc kubenswrapper[4559]: I1123 06:59:53.808763 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-597c64895-wx4bl" event={"ID":"94879e46-753f-4135-a501-c9ff00800450","Type":"ContainerStarted","Data":"2ebbc8b3578311e871c0068350a539927a2ecb40ed38dcc7b1d0c75cfafa82a5"} Nov 23 06:59:53 crc kubenswrapper[4559]: I1123 06:59:53.809284 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-597c64895-wx4bl" Nov 23 06:59:53 crc kubenswrapper[4559]: I1123 06:59:53.814993 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-55bb64d698-6lpkd" event={"ID":"f9264f09-464d-404a-88a1-7f86ecb2ccb8","Type":"ContainerStarted","Data":"687f658de34702c0b8cfc17d2d43ac5dd3ee7aa1edcae12107a023b633d18659"} Nov 23 06:59:53 crc kubenswrapper[4559]: I1123 06:59:53.815046 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-55bb64d698-6lpkd" event={"ID":"f9264f09-464d-404a-88a1-7f86ecb2ccb8","Type":"ContainerStarted","Data":"885b9105d35447b7d636888a150096108178af5f6741e52f07fc661f258c54c0"} Nov 23 06:59:53 crc kubenswrapper[4559]: I1123 06:59:53.815820 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-55bb64d698-6lpkd" Nov 23 06:59:53 crc kubenswrapper[4559]: I1123 06:59:53.815860 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-55bb64d698-6lpkd" Nov 23 06:59:53 crc kubenswrapper[4559]: I1123 06:59:53.857690 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-55bb64d698-6lpkd" podStartSLOduration=2.857674269 podStartE2EDuration="2.857674269s" podCreationTimestamp="2025-11-23 06:59:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:59:53.854331752 +0000 UTC m=+895.876317366" watchObservedRunningTime="2025-11-23 06:59:53.857674269 +0000 UTC m=+895.879659883" Nov 23 06:59:53 crc kubenswrapper[4559]: I1123 06:59:53.858745 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-597c64895-wx4bl" podStartSLOduration=4.858719154 podStartE2EDuration="4.858719154s" podCreationTimestamp="2025-11-23 06:59:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 06:59:53.828763876 +0000 UTC m=+895.850749490" watchObservedRunningTime="2025-11-23 06:59:53.858719154 +0000 UTC m=+895.880704768" Nov 23 06:59:54 crc kubenswrapper[4559]: I1123 06:59:54.827223 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-d4dd9b9cc-ndjpd" event={"ID":"f5b2ad32-159b-4369-951d-02d3ae8581eb","Type":"ContainerStarted","Data":"e42b0b5f9d610479a026b4442d2ff4a26aec66ae124a1364705a0d0ae21fba74"} Nov 23 06:59:54 crc kubenswrapper[4559]: I1123 06:59:54.829516 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-685df6659d-zwp5c" event={"ID":"cb7ba735-691f-4eb5-ac0c-c3f976e8e604","Type":"ContainerStarted","Data":"8c0885d3e8f07018471124a20915ab0720ef3bbbfeeef6def53ade7d16f5a69a"} Nov 23 06:59:54 crc kubenswrapper[4559]: I1123 06:59:54.832554 4559 generic.go:334] "Generic (PLEG): container finished" podID="c8ae1721-4ded-4e7d-9091-88ac908d0554" containerID="178d1c0f29794efd9880d96375c09387602578935cfa41dca7f541e19feea9ea" exitCode=0 Nov 23 06:59:54 crc kubenswrapper[4559]: I1123 06:59:54.833730 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-bxpjp" event={"ID":"c8ae1721-4ded-4e7d-9091-88ac908d0554","Type":"ContainerDied","Data":"178d1c0f29794efd9880d96375c09387602578935cfa41dca7f541e19feea9ea"} Nov 23 06:59:55 crc kubenswrapper[4559]: I1123 06:59:55.847800 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-d4dd9b9cc-ndjpd" event={"ID":"f5b2ad32-159b-4369-951d-02d3ae8581eb","Type":"ContainerStarted","Data":"d2e62184cb173d4534cb93d8108a7fe9c6d8c0db9d831275bbfe96f4de574f41"} Nov 23 06:59:55 crc kubenswrapper[4559]: I1123 06:59:55.861409 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-685df6659d-zwp5c" event={"ID":"cb7ba735-691f-4eb5-ac0c-c3f976e8e604","Type":"ContainerStarted","Data":"8b98188e14d898d6b6b060b1954fb3b3bc2c36e5fe8e3d8547819193cce01ba1"} Nov 23 06:59:55 crc kubenswrapper[4559]: I1123 06:59:55.882071 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-d4dd9b9cc-ndjpd" podStartSLOduration=4.189999437 podStartE2EDuration="6.882044484s" podCreationTimestamp="2025-11-23 06:59:49 +0000 UTC" firstStartedPulling="2025-11-23 06:59:51.881985428 +0000 UTC m=+893.903971042" lastFinishedPulling="2025-11-23 06:59:54.574030475 +0000 UTC m=+896.596016089" observedRunningTime="2025-11-23 06:59:55.870792393 +0000 UTC m=+897.892778007" watchObservedRunningTime="2025-11-23 06:59:55.882044484 +0000 UTC m=+897.904030099" Nov 23 06:59:55 crc kubenswrapper[4559]: I1123 06:59:55.893700 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-685df6659d-zwp5c" podStartSLOduration=4.075752452 podStartE2EDuration="6.893679367s" podCreationTimestamp="2025-11-23 06:59:49 +0000 UTC" firstStartedPulling="2025-11-23 06:59:51.753700119 +0000 UTC m=+893.775685734" lastFinishedPulling="2025-11-23 06:59:54.571627046 +0000 UTC m=+896.593612649" observedRunningTime="2025-11-23 06:59:55.886214588 +0000 UTC m=+897.908200202" watchObservedRunningTime="2025-11-23 06:59:55.893679367 +0000 UTC m=+897.915664981" Nov 23 06:59:56 crc kubenswrapper[4559]: I1123 06:59:56.167344 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 06:59:56 crc kubenswrapper[4559]: I1123 06:59:56.167421 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 06:59:57 crc kubenswrapper[4559]: I1123 06:59:57.995438 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-bxpjp" Nov 23 06:59:58 crc kubenswrapper[4559]: I1123 06:59:58.133893 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8ae1721-4ded-4e7d-9091-88ac908d0554-scripts\") pod \"c8ae1721-4ded-4e7d-9091-88ac908d0554\" (UID: \"c8ae1721-4ded-4e7d-9091-88ac908d0554\") " Nov 23 06:59:58 crc kubenswrapper[4559]: I1123 06:59:58.134053 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8ae1721-4ded-4e7d-9091-88ac908d0554-combined-ca-bundle\") pod \"c8ae1721-4ded-4e7d-9091-88ac908d0554\" (UID: \"c8ae1721-4ded-4e7d-9091-88ac908d0554\") " Nov 23 06:59:58 crc kubenswrapper[4559]: I1123 06:59:58.134078 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxzxp\" (UniqueName: \"kubernetes.io/projected/c8ae1721-4ded-4e7d-9091-88ac908d0554-kube-api-access-pxzxp\") pod \"c8ae1721-4ded-4e7d-9091-88ac908d0554\" (UID: \"c8ae1721-4ded-4e7d-9091-88ac908d0554\") " Nov 23 06:59:58 crc kubenswrapper[4559]: I1123 06:59:58.134167 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c8ae1721-4ded-4e7d-9091-88ac908d0554-etc-machine-id\") pod \"c8ae1721-4ded-4e7d-9091-88ac908d0554\" (UID: \"c8ae1721-4ded-4e7d-9091-88ac908d0554\") " Nov 23 06:59:58 crc kubenswrapper[4559]: I1123 06:59:58.134229 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c8ae1721-4ded-4e7d-9091-88ac908d0554-db-sync-config-data\") pod \"c8ae1721-4ded-4e7d-9091-88ac908d0554\" (UID: \"c8ae1721-4ded-4e7d-9091-88ac908d0554\") " Nov 23 06:59:58 crc kubenswrapper[4559]: I1123 06:59:58.134318 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8ae1721-4ded-4e7d-9091-88ac908d0554-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c8ae1721-4ded-4e7d-9091-88ac908d0554" (UID: "c8ae1721-4ded-4e7d-9091-88ac908d0554"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 06:59:58 crc kubenswrapper[4559]: I1123 06:59:58.134341 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8ae1721-4ded-4e7d-9091-88ac908d0554-config-data\") pod \"c8ae1721-4ded-4e7d-9091-88ac908d0554\" (UID: \"c8ae1721-4ded-4e7d-9091-88ac908d0554\") " Nov 23 06:59:58 crc kubenswrapper[4559]: I1123 06:59:58.135394 4559 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c8ae1721-4ded-4e7d-9091-88ac908d0554-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:58 crc kubenswrapper[4559]: I1123 06:59:58.139074 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8ae1721-4ded-4e7d-9091-88ac908d0554-kube-api-access-pxzxp" (OuterVolumeSpecName: "kube-api-access-pxzxp") pod "c8ae1721-4ded-4e7d-9091-88ac908d0554" (UID: "c8ae1721-4ded-4e7d-9091-88ac908d0554"). InnerVolumeSpecName "kube-api-access-pxzxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:59:58 crc kubenswrapper[4559]: I1123 06:59:58.139767 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8ae1721-4ded-4e7d-9091-88ac908d0554-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "c8ae1721-4ded-4e7d-9091-88ac908d0554" (UID: "c8ae1721-4ded-4e7d-9091-88ac908d0554"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:58 crc kubenswrapper[4559]: I1123 06:59:58.151445 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8ae1721-4ded-4e7d-9091-88ac908d0554-scripts" (OuterVolumeSpecName: "scripts") pod "c8ae1721-4ded-4e7d-9091-88ac908d0554" (UID: "c8ae1721-4ded-4e7d-9091-88ac908d0554"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:58 crc kubenswrapper[4559]: I1123 06:59:58.158751 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8ae1721-4ded-4e7d-9091-88ac908d0554-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c8ae1721-4ded-4e7d-9091-88ac908d0554" (UID: "c8ae1721-4ded-4e7d-9091-88ac908d0554"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:58 crc kubenswrapper[4559]: I1123 06:59:58.177390 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8ae1721-4ded-4e7d-9091-88ac908d0554-config-data" (OuterVolumeSpecName: "config-data") pod "c8ae1721-4ded-4e7d-9091-88ac908d0554" (UID: "c8ae1721-4ded-4e7d-9091-88ac908d0554"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 06:59:58 crc kubenswrapper[4559]: I1123 06:59:58.236850 4559 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c8ae1721-4ded-4e7d-9091-88ac908d0554-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:58 crc kubenswrapper[4559]: I1123 06:59:58.236881 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8ae1721-4ded-4e7d-9091-88ac908d0554-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:58 crc kubenswrapper[4559]: I1123 06:59:58.236891 4559 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8ae1721-4ded-4e7d-9091-88ac908d0554-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:58 crc kubenswrapper[4559]: I1123 06:59:58.236900 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8ae1721-4ded-4e7d-9091-88ac908d0554-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:58 crc kubenswrapper[4559]: I1123 06:59:58.236910 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxzxp\" (UniqueName: \"kubernetes.io/projected/c8ae1721-4ded-4e7d-9091-88ac908d0554-kube-api-access-pxzxp\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:58 crc kubenswrapper[4559]: I1123 06:59:58.888544 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-bxpjp" event={"ID":"c8ae1721-4ded-4e7d-9091-88ac908d0554","Type":"ContainerDied","Data":"6a9113470a7a2a0237d3f88cfdd768d72e140dfacc567addaaea7488009a56d6"} Nov 23 06:59:58 crc kubenswrapper[4559]: I1123 06:59:58.889041 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a9113470a7a2a0237d3f88cfdd768d72e140dfacc567addaaea7488009a56d6" Nov 23 06:59:58 crc kubenswrapper[4559]: I1123 06:59:58.888624 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-bxpjp" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.172757 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 06:59:59 crc kubenswrapper[4559]: E1123 06:59:59.173610 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8ae1721-4ded-4e7d-9091-88ac908d0554" containerName="cinder-db-sync" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.173628 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8ae1721-4ded-4e7d-9091-88ac908d0554" containerName="cinder-db-sync" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.173937 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8ae1721-4ded-4e7d-9091-88ac908d0554" containerName="cinder-db-sync" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.175874 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.182739 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.182986 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-7j26t" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.183150 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.183258 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.184303 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.255101 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-597c64895-wx4bl"] Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.255350 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-597c64895-wx4bl" podUID="94879e46-753f-4135-a501-c9ff00800450" containerName="dnsmasq-dns" containerID="cri-o://2ebbc8b3578311e871c0068350a539927a2ecb40ed38dcc7b1d0c75cfafa82a5" gracePeriod=10 Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.266452 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-597c64895-wx4bl" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.287950 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64cc7f6975-rdztg"] Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.289324 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.304481 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64cc7f6975-rdztg"] Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.369464 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mrcf\" (UniqueName: \"kubernetes.io/projected/cb5e8b5b-0513-4392-a079-0dab7664ec84-kube-api-access-2mrcf\") pod \"cinder-scheduler-0\" (UID: \"cb5e8b5b-0513-4392-a079-0dab7664ec84\") " pod="openstack/cinder-scheduler-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.369574 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb5e8b5b-0513-4392-a079-0dab7664ec84-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cb5e8b5b-0513-4392-a079-0dab7664ec84\") " pod="openstack/cinder-scheduler-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.369597 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb5e8b5b-0513-4392-a079-0dab7664ec84-scripts\") pod \"cinder-scheduler-0\" (UID: \"cb5e8b5b-0513-4392-a079-0dab7664ec84\") " pod="openstack/cinder-scheduler-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.369733 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb5e8b5b-0513-4392-a079-0dab7664ec84-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cb5e8b5b-0513-4392-a079-0dab7664ec84\") " pod="openstack/cinder-scheduler-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.369773 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb5e8b5b-0513-4392-a079-0dab7664ec84-config-data\") pod \"cinder-scheduler-0\" (UID: \"cb5e8b5b-0513-4392-a079-0dab7664ec84\") " pod="openstack/cinder-scheduler-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.369798 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb5e8b5b-0513-4392-a079-0dab7664ec84-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cb5e8b5b-0513-4392-a079-0dab7664ec84\") " pod="openstack/cinder-scheduler-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.423697 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.425159 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.427320 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.430321 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.471937 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb5e8b5b-0513-4392-a079-0dab7664ec84-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cb5e8b5b-0513-4392-a079-0dab7664ec84\") " pod="openstack/cinder-scheduler-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.471977 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-dns-svc\") pod \"dnsmasq-dns-64cc7f6975-rdztg\" (UID: \"d6acba84-09d4-4657-8c25-b863919294a2\") " pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.472023 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb5e8b5b-0513-4392-a079-0dab7664ec84-config-data\") pod \"cinder-scheduler-0\" (UID: \"cb5e8b5b-0513-4392-a079-0dab7664ec84\") " pod="openstack/cinder-scheduler-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.472051 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-ovsdbserver-nb\") pod \"dnsmasq-dns-64cc7f6975-rdztg\" (UID: \"d6acba84-09d4-4657-8c25-b863919294a2\") " pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.472072 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb5e8b5b-0513-4392-a079-0dab7664ec84-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cb5e8b5b-0513-4392-a079-0dab7664ec84\") " pod="openstack/cinder-scheduler-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.472135 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mrcf\" (UniqueName: \"kubernetes.io/projected/cb5e8b5b-0513-4392-a079-0dab7664ec84-kube-api-access-2mrcf\") pod \"cinder-scheduler-0\" (UID: \"cb5e8b5b-0513-4392-a079-0dab7664ec84\") " pod="openstack/cinder-scheduler-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.472162 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5dgf\" (UniqueName: \"kubernetes.io/projected/d6acba84-09d4-4657-8c25-b863919294a2-kube-api-access-d5dgf\") pod \"dnsmasq-dns-64cc7f6975-rdztg\" (UID: \"d6acba84-09d4-4657-8c25-b863919294a2\") " pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.472187 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-dns-swift-storage-0\") pod \"dnsmasq-dns-64cc7f6975-rdztg\" (UID: \"d6acba84-09d4-4657-8c25-b863919294a2\") " pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.472248 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb5e8b5b-0513-4392-a079-0dab7664ec84-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cb5e8b5b-0513-4392-a079-0dab7664ec84\") " pod="openstack/cinder-scheduler-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.472266 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb5e8b5b-0513-4392-a079-0dab7664ec84-scripts\") pod \"cinder-scheduler-0\" (UID: \"cb5e8b5b-0513-4392-a079-0dab7664ec84\") " pod="openstack/cinder-scheduler-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.472330 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-ovsdbserver-sb\") pod \"dnsmasq-dns-64cc7f6975-rdztg\" (UID: \"d6acba84-09d4-4657-8c25-b863919294a2\") " pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.472356 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-config\") pod \"dnsmasq-dns-64cc7f6975-rdztg\" (UID: \"d6acba84-09d4-4657-8c25-b863919294a2\") " pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.475061 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb5e8b5b-0513-4392-a079-0dab7664ec84-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cb5e8b5b-0513-4392-a079-0dab7664ec84\") " pod="openstack/cinder-scheduler-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.479443 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb5e8b5b-0513-4392-a079-0dab7664ec84-scripts\") pod \"cinder-scheduler-0\" (UID: \"cb5e8b5b-0513-4392-a079-0dab7664ec84\") " pod="openstack/cinder-scheduler-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.479742 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb5e8b5b-0513-4392-a079-0dab7664ec84-config-data\") pod \"cinder-scheduler-0\" (UID: \"cb5e8b5b-0513-4392-a079-0dab7664ec84\") " pod="openstack/cinder-scheduler-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.480124 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb5e8b5b-0513-4392-a079-0dab7664ec84-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cb5e8b5b-0513-4392-a079-0dab7664ec84\") " pod="openstack/cinder-scheduler-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.481107 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb5e8b5b-0513-4392-a079-0dab7664ec84-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cb5e8b5b-0513-4392-a079-0dab7664ec84\") " pod="openstack/cinder-scheduler-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.498419 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mrcf\" (UniqueName: \"kubernetes.io/projected/cb5e8b5b-0513-4392-a079-0dab7664ec84-kube-api-access-2mrcf\") pod \"cinder-scheduler-0\" (UID: \"cb5e8b5b-0513-4392-a079-0dab7664ec84\") " pod="openstack/cinder-scheduler-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.546964 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.574194 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-dns-svc\") pod \"dnsmasq-dns-64cc7f6975-rdztg\" (UID: \"d6acba84-09d4-4657-8c25-b863919294a2\") " pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.574278 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-ovsdbserver-nb\") pod \"dnsmasq-dns-64cc7f6975-rdztg\" (UID: \"d6acba84-09d4-4657-8c25-b863919294a2\") " pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.574382 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5dgf\" (UniqueName: \"kubernetes.io/projected/d6acba84-09d4-4657-8c25-b863919294a2-kube-api-access-d5dgf\") pod \"dnsmasq-dns-64cc7f6975-rdztg\" (UID: \"d6acba84-09d4-4657-8c25-b863919294a2\") " pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.574411 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-dns-swift-storage-0\") pod \"dnsmasq-dns-64cc7f6975-rdztg\" (UID: \"d6acba84-09d4-4657-8c25-b863919294a2\") " pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.574444 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-config-data-custom\") pod \"cinder-api-0\" (UID: \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\") " pod="openstack/cinder-api-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.574492 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\") " pod="openstack/cinder-api-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.574560 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-ovsdbserver-sb\") pod \"dnsmasq-dns-64cc7f6975-rdztg\" (UID: \"d6acba84-09d4-4657-8c25-b863919294a2\") " pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.574593 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-828lk\" (UniqueName: \"kubernetes.io/projected/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-kube-api-access-828lk\") pod \"cinder-api-0\" (UID: \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\") " pod="openstack/cinder-api-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.574617 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-config\") pod \"dnsmasq-dns-64cc7f6975-rdztg\" (UID: \"d6acba84-09d4-4657-8c25-b863919294a2\") " pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.574666 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-logs\") pod \"cinder-api-0\" (UID: \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\") " pod="openstack/cinder-api-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.574683 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-scripts\") pod \"cinder-api-0\" (UID: \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\") " pod="openstack/cinder-api-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.574725 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-config-data\") pod \"cinder-api-0\" (UID: \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\") " pod="openstack/cinder-api-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.574762 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\") " pod="openstack/cinder-api-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.575824 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-dns-svc\") pod \"dnsmasq-dns-64cc7f6975-rdztg\" (UID: \"d6acba84-09d4-4657-8c25-b863919294a2\") " pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.576073 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-ovsdbserver-nb\") pod \"dnsmasq-dns-64cc7f6975-rdztg\" (UID: \"d6acba84-09d4-4657-8c25-b863919294a2\") " pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.576407 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-config\") pod \"dnsmasq-dns-64cc7f6975-rdztg\" (UID: \"d6acba84-09d4-4657-8c25-b863919294a2\") " pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.577357 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-dns-swift-storage-0\") pod \"dnsmasq-dns-64cc7f6975-rdztg\" (UID: \"d6acba84-09d4-4657-8c25-b863919294a2\") " pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.577816 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-ovsdbserver-sb\") pod \"dnsmasq-dns-64cc7f6975-rdztg\" (UID: \"d6acba84-09d4-4657-8c25-b863919294a2\") " pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.592427 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5dgf\" (UniqueName: \"kubernetes.io/projected/d6acba84-09d4-4657-8c25-b863919294a2-kube-api-access-d5dgf\") pod \"dnsmasq-dns-64cc7f6975-rdztg\" (UID: \"d6acba84-09d4-4657-8c25-b863919294a2\") " pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.608364 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.676836 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-828lk\" (UniqueName: \"kubernetes.io/projected/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-kube-api-access-828lk\") pod \"cinder-api-0\" (UID: \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\") " pod="openstack/cinder-api-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.676892 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-logs\") pod \"cinder-api-0\" (UID: \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\") " pod="openstack/cinder-api-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.676910 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-scripts\") pod \"cinder-api-0\" (UID: \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\") " pod="openstack/cinder-api-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.676960 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-config-data\") pod \"cinder-api-0\" (UID: \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\") " pod="openstack/cinder-api-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.676995 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\") " pod="openstack/cinder-api-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.677167 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-config-data-custom\") pod \"cinder-api-0\" (UID: \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\") " pod="openstack/cinder-api-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.677212 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\") " pod="openstack/cinder-api-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.678397 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\") " pod="openstack/cinder-api-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.680358 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-logs\") pod \"cinder-api-0\" (UID: \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\") " pod="openstack/cinder-api-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.685620 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-config-data\") pod \"cinder-api-0\" (UID: \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\") " pod="openstack/cinder-api-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.686023 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\") " pod="openstack/cinder-api-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.686099 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-scripts\") pod \"cinder-api-0\" (UID: \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\") " pod="openstack/cinder-api-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.686132 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-config-data-custom\") pod \"cinder-api-0\" (UID: \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\") " pod="openstack/cinder-api-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.696128 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-828lk\" (UniqueName: \"kubernetes.io/projected/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-kube-api-access-828lk\") pod \"cinder-api-0\" (UID: \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\") " pod="openstack/cinder-api-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.703244 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-597c64895-wx4bl" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.754281 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.881827 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-dns-svc\") pod \"94879e46-753f-4135-a501-c9ff00800450\" (UID: \"94879e46-753f-4135-a501-c9ff00800450\") " Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.881972 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dklj7\" (UniqueName: \"kubernetes.io/projected/94879e46-753f-4135-a501-c9ff00800450-kube-api-access-dklj7\") pod \"94879e46-753f-4135-a501-c9ff00800450\" (UID: \"94879e46-753f-4135-a501-c9ff00800450\") " Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.882018 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-ovsdbserver-sb\") pod \"94879e46-753f-4135-a501-c9ff00800450\" (UID: \"94879e46-753f-4135-a501-c9ff00800450\") " Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.882160 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-config\") pod \"94879e46-753f-4135-a501-c9ff00800450\" (UID: \"94879e46-753f-4135-a501-c9ff00800450\") " Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.882215 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-dns-swift-storage-0\") pod \"94879e46-753f-4135-a501-c9ff00800450\" (UID: \"94879e46-753f-4135-a501-c9ff00800450\") " Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.882244 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-ovsdbserver-nb\") pod \"94879e46-753f-4135-a501-c9ff00800450\" (UID: \"94879e46-753f-4135-a501-c9ff00800450\") " Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.912950 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94879e46-753f-4135-a501-c9ff00800450-kube-api-access-dklj7" (OuterVolumeSpecName: "kube-api-access-dklj7") pod "94879e46-753f-4135-a501-c9ff00800450" (UID: "94879e46-753f-4135-a501-c9ff00800450"). InnerVolumeSpecName "kube-api-access-dklj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.933825 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "94879e46-753f-4135-a501-c9ff00800450" (UID: "94879e46-753f-4135-a501-c9ff00800450"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.933962 4559 generic.go:334] "Generic (PLEG): container finished" podID="94879e46-753f-4135-a501-c9ff00800450" containerID="2ebbc8b3578311e871c0068350a539927a2ecb40ed38dcc7b1d0c75cfafa82a5" exitCode=0 Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.934039 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-597c64895-wx4bl" event={"ID":"94879e46-753f-4135-a501-c9ff00800450","Type":"ContainerDied","Data":"2ebbc8b3578311e871c0068350a539927a2ecb40ed38dcc7b1d0c75cfafa82a5"} Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.934097 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-597c64895-wx4bl" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.934118 4559 scope.go:117] "RemoveContainer" containerID="2ebbc8b3578311e871c0068350a539927a2ecb40ed38dcc7b1d0c75cfafa82a5" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.934104 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-597c64895-wx4bl" event={"ID":"94879e46-753f-4135-a501-c9ff00800450","Type":"ContainerDied","Data":"a315c7cb9f283bb090318c6799ff819ff6be2ddb5350db39d4500ecb1aed3292"} Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.938121 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "94879e46-753f-4135-a501-c9ff00800450" (UID: "94879e46-753f-4135-a501-c9ff00800450"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.939994 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-config" (OuterVolumeSpecName: "config") pod "94879e46-753f-4135-a501-c9ff00800450" (UID: "94879e46-753f-4135-a501-c9ff00800450"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.941087 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "94879e46-753f-4135-a501-c9ff00800450" (UID: "94879e46-753f-4135-a501-c9ff00800450"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.942257 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4163bfe8-0de7-4490-b825-55d0f0343787","Type":"ContainerStarted","Data":"45189c2b7acb1f8b18c8954aedcd433313f126fd67dde19896cb9e5e255dee02"} Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.942396 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.942392 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4163bfe8-0de7-4490-b825-55d0f0343787" containerName="ceilometer-central-agent" containerID="cri-o://ba10afde1644658d448c1bb320ea05c40fdfb8995a31b98a833593e1267bbfbe" gracePeriod=30 Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.942537 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4163bfe8-0de7-4490-b825-55d0f0343787" containerName="proxy-httpd" containerID="cri-o://45189c2b7acb1f8b18c8954aedcd433313f126fd67dde19896cb9e5e255dee02" gracePeriod=30 Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.942635 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4163bfe8-0de7-4490-b825-55d0f0343787" containerName="sg-core" containerID="cri-o://5fb7c20c88f83d97306e5172d980cc901c730345621912965ed251efeb54307a" gracePeriod=30 Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.942714 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4163bfe8-0de7-4490-b825-55d0f0343787" containerName="ceilometer-notification-agent" containerID="cri-o://cd785a0026c59d5e74c27422561bc4c114f548239eeb6189968a2dd70cd1543d" gracePeriod=30 Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.953160 4559 scope.go:117] "RemoveContainer" containerID="e505ab11f0d0748ac9c3d7c30dfcda70af23e0a4d8b39029599104affd2c85ef" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.956421 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "94879e46-753f-4135-a501-c9ff00800450" (UID: "94879e46-753f-4135-a501-c9ff00800450"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.969421 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.914305244 podStartE2EDuration="34.969397348s" podCreationTimestamp="2025-11-23 06:59:25 +0000 UTC" firstStartedPulling="2025-11-23 06:59:26.019827461 +0000 UTC m=+868.041813074" lastFinishedPulling="2025-11-23 06:59:59.074919564 +0000 UTC m=+901.096905178" observedRunningTime="2025-11-23 06:59:59.962912381 +0000 UTC m=+901.984897995" watchObservedRunningTime="2025-11-23 06:59:59.969397348 +0000 UTC m=+901.991382961" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.980963 4559 scope.go:117] "RemoveContainer" containerID="2ebbc8b3578311e871c0068350a539927a2ecb40ed38dcc7b1d0c75cfafa82a5" Nov 23 06:59:59 crc kubenswrapper[4559]: E1123 06:59:59.981383 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ebbc8b3578311e871c0068350a539927a2ecb40ed38dcc7b1d0c75cfafa82a5\": container with ID starting with 2ebbc8b3578311e871c0068350a539927a2ecb40ed38dcc7b1d0c75cfafa82a5 not found: ID does not exist" containerID="2ebbc8b3578311e871c0068350a539927a2ecb40ed38dcc7b1d0c75cfafa82a5" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.981429 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ebbc8b3578311e871c0068350a539927a2ecb40ed38dcc7b1d0c75cfafa82a5"} err="failed to get container status \"2ebbc8b3578311e871c0068350a539927a2ecb40ed38dcc7b1d0c75cfafa82a5\": rpc error: code = NotFound desc = could not find container \"2ebbc8b3578311e871c0068350a539927a2ecb40ed38dcc7b1d0c75cfafa82a5\": container with ID starting with 2ebbc8b3578311e871c0068350a539927a2ecb40ed38dcc7b1d0c75cfafa82a5 not found: ID does not exist" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.981460 4559 scope.go:117] "RemoveContainer" containerID="e505ab11f0d0748ac9c3d7c30dfcda70af23e0a4d8b39029599104affd2c85ef" Nov 23 06:59:59 crc kubenswrapper[4559]: E1123 06:59:59.981857 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e505ab11f0d0748ac9c3d7c30dfcda70af23e0a4d8b39029599104affd2c85ef\": container with ID starting with e505ab11f0d0748ac9c3d7c30dfcda70af23e0a4d8b39029599104affd2c85ef not found: ID does not exist" containerID="e505ab11f0d0748ac9c3d7c30dfcda70af23e0a4d8b39029599104affd2c85ef" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.981898 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e505ab11f0d0748ac9c3d7c30dfcda70af23e0a4d8b39029599104affd2c85ef"} err="failed to get container status \"e505ab11f0d0748ac9c3d7c30dfcda70af23e0a4d8b39029599104affd2c85ef\": rpc error: code = NotFound desc = could not find container \"e505ab11f0d0748ac9c3d7c30dfcda70af23e0a4d8b39029599104affd2c85ef\": container with ID starting with e505ab11f0d0748ac9c3d7c30dfcda70af23e0a4d8b39029599104affd2c85ef not found: ID does not exist" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.984933 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dklj7\" (UniqueName: \"kubernetes.io/projected/94879e46-753f-4135-a501-c9ff00800450-kube-api-access-dklj7\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.984958 4559 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.984969 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-config\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.984978 4559 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.984990 4559 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 06:59:59 crc kubenswrapper[4559]: I1123 06:59:59.984998 4559 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94879e46-753f-4135-a501-c9ff00800450-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.061670 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 07:00:00 crc kubenswrapper[4559]: W1123 07:00:00.067188 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb5e8b5b_0513_4392_a079_0dab7664ec84.slice/crio-17e7bb197c8a3788374cdec264eb06fb2f9ec9e587d6787d61eb898d47efa22a WatchSource:0}: Error finding container 17e7bb197c8a3788374cdec264eb06fb2f9ec9e587d6787d61eb898d47efa22a: Status 404 returned error can't find the container with id 17e7bb197c8a3788374cdec264eb06fb2f9ec9e587d6787d61eb898d47efa22a Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.127873 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398020-nvnnf"] Nov 23 07:00:00 crc kubenswrapper[4559]: E1123 07:00:00.128430 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94879e46-753f-4135-a501-c9ff00800450" containerName="init" Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.128449 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="94879e46-753f-4135-a501-c9ff00800450" containerName="init" Nov 23 07:00:00 crc kubenswrapper[4559]: E1123 07:00:00.128462 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94879e46-753f-4135-a501-c9ff00800450" containerName="dnsmasq-dns" Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.128468 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="94879e46-753f-4135-a501-c9ff00800450" containerName="dnsmasq-dns" Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.128697 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="94879e46-753f-4135-a501-c9ff00800450" containerName="dnsmasq-dns" Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.129560 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-nvnnf" Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.133635 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.133865 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.154327 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398020-nvnnf"] Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.236416 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64cc7f6975-rdztg"] Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.247203 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.311413 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b5ec40f-303a-4c84-bf2a-8bf8991974df-secret-volume\") pod \"collect-profiles-29398020-nvnnf\" (UID: \"5b5ec40f-303a-4c84-bf2a-8bf8991974df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-nvnnf" Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.312026 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b5ec40f-303a-4c84-bf2a-8bf8991974df-config-volume\") pod \"collect-profiles-29398020-nvnnf\" (UID: \"5b5ec40f-303a-4c84-bf2a-8bf8991974df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-nvnnf" Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.312080 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plhvv\" (UniqueName: \"kubernetes.io/projected/5b5ec40f-303a-4c84-bf2a-8bf8991974df-kube-api-access-plhvv\") pod \"collect-profiles-29398020-nvnnf\" (UID: \"5b5ec40f-303a-4c84-bf2a-8bf8991974df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-nvnnf" Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.348851 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-597c64895-wx4bl"] Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.357622 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-597c64895-wx4bl"] Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.417032 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b5ec40f-303a-4c84-bf2a-8bf8991974df-secret-volume\") pod \"collect-profiles-29398020-nvnnf\" (UID: \"5b5ec40f-303a-4c84-bf2a-8bf8991974df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-nvnnf" Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.418083 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b5ec40f-303a-4c84-bf2a-8bf8991974df-config-volume\") pod \"collect-profiles-29398020-nvnnf\" (UID: \"5b5ec40f-303a-4c84-bf2a-8bf8991974df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-nvnnf" Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.418114 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plhvv\" (UniqueName: \"kubernetes.io/projected/5b5ec40f-303a-4c84-bf2a-8bf8991974df-kube-api-access-plhvv\") pod \"collect-profiles-29398020-nvnnf\" (UID: \"5b5ec40f-303a-4c84-bf2a-8bf8991974df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-nvnnf" Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.420744 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b5ec40f-303a-4c84-bf2a-8bf8991974df-config-volume\") pod \"collect-profiles-29398020-nvnnf\" (UID: \"5b5ec40f-303a-4c84-bf2a-8bf8991974df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-nvnnf" Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.423105 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b5ec40f-303a-4c84-bf2a-8bf8991974df-secret-volume\") pod \"collect-profiles-29398020-nvnnf\" (UID: \"5b5ec40f-303a-4c84-bf2a-8bf8991974df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-nvnnf" Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.434967 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plhvv\" (UniqueName: \"kubernetes.io/projected/5b5ec40f-303a-4c84-bf2a-8bf8991974df-kube-api-access-plhvv\") pod \"collect-profiles-29398020-nvnnf\" (UID: \"5b5ec40f-303a-4c84-bf2a-8bf8991974df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-nvnnf" Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.500216 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-nvnnf" Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.934950 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398020-nvnnf"] Nov 23 07:00:00 crc kubenswrapper[4559]: W1123 07:00:00.945603 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b5ec40f_303a_4c84_bf2a_8bf8991974df.slice/crio-2355633c5f2024a7a843ff258274db073c1e237924af9c433b08f0b06b6d0dd9 WatchSource:0}: Error finding container 2355633c5f2024a7a843ff258274db073c1e237924af9c433b08f0b06b6d0dd9: Status 404 returned error can't find the container with id 2355633c5f2024a7a843ff258274db073c1e237924af9c433b08f0b06b6d0dd9 Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.959173 4559 generic.go:334] "Generic (PLEG): container finished" podID="d6acba84-09d4-4657-8c25-b863919294a2" containerID="401abf9d16ce3ec67e1fcbd50bb57d6e050b1ae263b46da20fbb84f857f9bb65" exitCode=0 Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.959242 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" event={"ID":"d6acba84-09d4-4657-8c25-b863919294a2","Type":"ContainerDied","Data":"401abf9d16ce3ec67e1fcbd50bb57d6e050b1ae263b46da20fbb84f857f9bb65"} Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.959295 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" event={"ID":"d6acba84-09d4-4657-8c25-b863919294a2","Type":"ContainerStarted","Data":"6ef48fdd1d3bfc863106c8e1d3c7fbe097bc254d1d11a662310b29bac0e63730"} Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.982538 4559 generic.go:334] "Generic (PLEG): container finished" podID="4163bfe8-0de7-4490-b825-55d0f0343787" containerID="45189c2b7acb1f8b18c8954aedcd433313f126fd67dde19896cb9e5e255dee02" exitCode=0 Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.982568 4559 generic.go:334] "Generic (PLEG): container finished" podID="4163bfe8-0de7-4490-b825-55d0f0343787" containerID="5fb7c20c88f83d97306e5172d980cc901c730345621912965ed251efeb54307a" exitCode=2 Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.982576 4559 generic.go:334] "Generic (PLEG): container finished" podID="4163bfe8-0de7-4490-b825-55d0f0343787" containerID="ba10afde1644658d448c1bb320ea05c40fdfb8995a31b98a833593e1267bbfbe" exitCode=0 Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.982628 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4163bfe8-0de7-4490-b825-55d0f0343787","Type":"ContainerDied","Data":"45189c2b7acb1f8b18c8954aedcd433313f126fd67dde19896cb9e5e255dee02"} Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.982678 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4163bfe8-0de7-4490-b825-55d0f0343787","Type":"ContainerDied","Data":"5fb7c20c88f83d97306e5172d980cc901c730345621912965ed251efeb54307a"} Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.982691 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4163bfe8-0de7-4490-b825-55d0f0343787","Type":"ContainerDied","Data":"ba10afde1644658d448c1bb320ea05c40fdfb8995a31b98a833593e1267bbfbe"} Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.989758 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"aec4f5ef-bdf5-475a-8a2b-5d70751161c6","Type":"ContainerStarted","Data":"a75e75394bad505488d6559266c3259e1e947ffca421dc069ad241dde82a7f95"} Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.989799 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"aec4f5ef-bdf5-475a-8a2b-5d70751161c6","Type":"ContainerStarted","Data":"39e46bbbeb01a8eacf83c00669229ce1311b92e8a135c71c150726cb3d760342"} Nov 23 07:00:00 crc kubenswrapper[4559]: I1123 07:00:00.992992 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb5e8b5b-0513-4392-a079-0dab7664ec84","Type":"ContainerStarted","Data":"17e7bb197c8a3788374cdec264eb06fb2f9ec9e587d6787d61eb898d47efa22a"} Nov 23 07:00:01 crc kubenswrapper[4559]: I1123 07:00:01.162370 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 23 07:00:01 crc kubenswrapper[4559]: I1123 07:00:01.593976 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-798c5546f8-28xxs" Nov 23 07:00:01 crc kubenswrapper[4559]: I1123 07:00:01.681317 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-798c5546f8-28xxs" Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.009354 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb5e8b5b-0513-4392-a079-0dab7664ec84","Type":"ContainerStarted","Data":"1d0ef4f29080e8dba7e9ab6b9fc27101b075c4bc0291cf4f4b8f734983e8543a"} Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.011436 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" event={"ID":"d6acba84-09d4-4657-8c25-b863919294a2","Type":"ContainerStarted","Data":"b6e390e295765cffce68ec1cc51febc51204bd83a3fd7e2e7ea92c138ecbaa16"} Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.012426 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.014549 4559 generic.go:334] "Generic (PLEG): container finished" podID="5b5ec40f-303a-4c84-bf2a-8bf8991974df" containerID="f2a3dcb5f3977c7549a38eac6d400b53cc317877dd08e94cde33bf3c041a5a54" exitCode=0 Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.014591 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-nvnnf" event={"ID":"5b5ec40f-303a-4c84-bf2a-8bf8991974df","Type":"ContainerDied","Data":"f2a3dcb5f3977c7549a38eac6d400b53cc317877dd08e94cde33bf3c041a5a54"} Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.014607 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-nvnnf" event={"ID":"5b5ec40f-303a-4c84-bf2a-8bf8991974df","Type":"ContainerStarted","Data":"2355633c5f2024a7a843ff258274db073c1e237924af9c433b08f0b06b6d0dd9"} Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.023058 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="aec4f5ef-bdf5-475a-8a2b-5d70751161c6" containerName="cinder-api-log" containerID="cri-o://a75e75394bad505488d6559266c3259e1e947ffca421dc069ad241dde82a7f95" gracePeriod=30 Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.023122 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"aec4f5ef-bdf5-475a-8a2b-5d70751161c6","Type":"ContainerStarted","Data":"46b8f2b2ff14784cbf9d6c17a44eb00e0191943703d04d3d075c37adec30ec5d"} Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.023147 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.023171 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="aec4f5ef-bdf5-475a-8a2b-5d70751161c6" containerName="cinder-api" containerID="cri-o://46b8f2b2ff14784cbf9d6c17a44eb00e0191943703d04d3d075c37adec30ec5d" gracePeriod=30 Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.046887 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" podStartSLOduration=3.046874373 podStartE2EDuration="3.046874373s" podCreationTimestamp="2025-11-23 06:59:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:00:02.030539171 +0000 UTC m=+904.052524785" watchObservedRunningTime="2025-11-23 07:00:02.046874373 +0000 UTC m=+904.068859987" Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.050907 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.050897801 podStartE2EDuration="3.050897801s" podCreationTimestamp="2025-11-23 06:59:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:00:02.045227346 +0000 UTC m=+904.067212960" watchObservedRunningTime="2025-11-23 07:00:02.050897801 +0000 UTC m=+904.072883415" Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.283904 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94879e46-753f-4135-a501-c9ff00800450" path="/var/lib/kubelet/pods/94879e46-753f-4135-a501-c9ff00800450/volumes" Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.602300 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.676826 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-828lk\" (UniqueName: \"kubernetes.io/projected/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-kube-api-access-828lk\") pod \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\" (UID: \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\") " Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.676899 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-logs\") pod \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\" (UID: \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\") " Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.676961 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-config-data\") pod \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\" (UID: \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\") " Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.677030 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-config-data-custom\") pod \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\" (UID: \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\") " Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.677083 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-combined-ca-bundle\") pod \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\" (UID: \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\") " Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.677140 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-etc-machine-id\") pod \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\" (UID: \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\") " Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.677189 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-scripts\") pod \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\" (UID: \"aec4f5ef-bdf5-475a-8a2b-5d70751161c6\") " Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.680985 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-logs" (OuterVolumeSpecName: "logs") pod "aec4f5ef-bdf5-475a-8a2b-5d70751161c6" (UID: "aec4f5ef-bdf5-475a-8a2b-5d70751161c6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.681526 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "aec4f5ef-bdf5-475a-8a2b-5d70751161c6" (UID: "aec4f5ef-bdf5-475a-8a2b-5d70751161c6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.704763 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "aec4f5ef-bdf5-475a-8a2b-5d70751161c6" (UID: "aec4f5ef-bdf5-475a-8a2b-5d70751161c6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.706791 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-kube-api-access-828lk" (OuterVolumeSpecName: "kube-api-access-828lk") pod "aec4f5ef-bdf5-475a-8a2b-5d70751161c6" (UID: "aec4f5ef-bdf5-475a-8a2b-5d70751161c6"). InnerVolumeSpecName "kube-api-access-828lk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.721729 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-scripts" (OuterVolumeSpecName: "scripts") pod "aec4f5ef-bdf5-475a-8a2b-5d70751161c6" (UID: "aec4f5ef-bdf5-475a-8a2b-5d70751161c6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.723740 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aec4f5ef-bdf5-475a-8a2b-5d70751161c6" (UID: "aec4f5ef-bdf5-475a-8a2b-5d70751161c6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.735787 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-config-data" (OuterVolumeSpecName: "config-data") pod "aec4f5ef-bdf5-475a-8a2b-5d70751161c6" (UID: "aec4f5ef-bdf5-475a-8a2b-5d70751161c6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.781135 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.781167 4559 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.781179 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.781188 4559 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.781198 4559 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.781206 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-828lk\" (UniqueName: \"kubernetes.io/projected/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-kube-api-access-828lk\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:02 crc kubenswrapper[4559]: I1123 07:00:02.781215 4559 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aec4f5ef-bdf5-475a-8a2b-5d70751161c6-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.035166 4559 generic.go:334] "Generic (PLEG): container finished" podID="aec4f5ef-bdf5-475a-8a2b-5d70751161c6" containerID="46b8f2b2ff14784cbf9d6c17a44eb00e0191943703d04d3d075c37adec30ec5d" exitCode=0 Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.035424 4559 generic.go:334] "Generic (PLEG): container finished" podID="aec4f5ef-bdf5-475a-8a2b-5d70751161c6" containerID="a75e75394bad505488d6559266c3259e1e947ffca421dc069ad241dde82a7f95" exitCode=143 Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.035220 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.035239 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"aec4f5ef-bdf5-475a-8a2b-5d70751161c6","Type":"ContainerDied","Data":"46b8f2b2ff14784cbf9d6c17a44eb00e0191943703d04d3d075c37adec30ec5d"} Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.035511 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"aec4f5ef-bdf5-475a-8a2b-5d70751161c6","Type":"ContainerDied","Data":"a75e75394bad505488d6559266c3259e1e947ffca421dc069ad241dde82a7f95"} Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.035525 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"aec4f5ef-bdf5-475a-8a2b-5d70751161c6","Type":"ContainerDied","Data":"39e46bbbeb01a8eacf83c00669229ce1311b92e8a135c71c150726cb3d760342"} Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.035543 4559 scope.go:117] "RemoveContainer" containerID="46b8f2b2ff14784cbf9d6c17a44eb00e0191943703d04d3d075c37adec30ec5d" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.037541 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb5e8b5b-0513-4392-a079-0dab7664ec84","Type":"ContainerStarted","Data":"704862b0ff2fe684d984c817f5989629201f75ed385a420513da5c764edd5a15"} Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.072062 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.180004586 podStartE2EDuration="4.07204785s" podCreationTimestamp="2025-11-23 06:59:59 +0000 UTC" firstStartedPulling="2025-11-23 07:00:00.070541061 +0000 UTC m=+902.092526676" lastFinishedPulling="2025-11-23 07:00:00.962584325 +0000 UTC m=+902.984569940" observedRunningTime="2025-11-23 07:00:03.061921124 +0000 UTC m=+905.083906738" watchObservedRunningTime="2025-11-23 07:00:03.07204785 +0000 UTC m=+905.094033463" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.076696 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.080125 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.090796 4559 scope.go:117] "RemoveContainer" containerID="a75e75394bad505488d6559266c3259e1e947ffca421dc069ad241dde82a7f95" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.097700 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 23 07:00:03 crc kubenswrapper[4559]: E1123 07:00:03.098081 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aec4f5ef-bdf5-475a-8a2b-5d70751161c6" containerName="cinder-api-log" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.098099 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="aec4f5ef-bdf5-475a-8a2b-5d70751161c6" containerName="cinder-api-log" Nov 23 07:00:03 crc kubenswrapper[4559]: E1123 07:00:03.098138 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aec4f5ef-bdf5-475a-8a2b-5d70751161c6" containerName="cinder-api" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.098144 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="aec4f5ef-bdf5-475a-8a2b-5d70751161c6" containerName="cinder-api" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.098312 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="aec4f5ef-bdf5-475a-8a2b-5d70751161c6" containerName="cinder-api-log" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.098335 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="aec4f5ef-bdf5-475a-8a2b-5d70751161c6" containerName="cinder-api" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.099231 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.104141 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.104273 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.104370 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.115768 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.142656 4559 scope.go:117] "RemoveContainer" containerID="46b8f2b2ff14784cbf9d6c17a44eb00e0191943703d04d3d075c37adec30ec5d" Nov 23 07:00:03 crc kubenswrapper[4559]: E1123 07:00:03.144118 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46b8f2b2ff14784cbf9d6c17a44eb00e0191943703d04d3d075c37adec30ec5d\": container with ID starting with 46b8f2b2ff14784cbf9d6c17a44eb00e0191943703d04d3d075c37adec30ec5d not found: ID does not exist" containerID="46b8f2b2ff14784cbf9d6c17a44eb00e0191943703d04d3d075c37adec30ec5d" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.144166 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46b8f2b2ff14784cbf9d6c17a44eb00e0191943703d04d3d075c37adec30ec5d"} err="failed to get container status \"46b8f2b2ff14784cbf9d6c17a44eb00e0191943703d04d3d075c37adec30ec5d\": rpc error: code = NotFound desc = could not find container \"46b8f2b2ff14784cbf9d6c17a44eb00e0191943703d04d3d075c37adec30ec5d\": container with ID starting with 46b8f2b2ff14784cbf9d6c17a44eb00e0191943703d04d3d075c37adec30ec5d not found: ID does not exist" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.144219 4559 scope.go:117] "RemoveContainer" containerID="a75e75394bad505488d6559266c3259e1e947ffca421dc069ad241dde82a7f95" Nov 23 07:00:03 crc kubenswrapper[4559]: E1123 07:00:03.145501 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a75e75394bad505488d6559266c3259e1e947ffca421dc069ad241dde82a7f95\": container with ID starting with a75e75394bad505488d6559266c3259e1e947ffca421dc069ad241dde82a7f95 not found: ID does not exist" containerID="a75e75394bad505488d6559266c3259e1e947ffca421dc069ad241dde82a7f95" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.145551 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a75e75394bad505488d6559266c3259e1e947ffca421dc069ad241dde82a7f95"} err="failed to get container status \"a75e75394bad505488d6559266c3259e1e947ffca421dc069ad241dde82a7f95\": rpc error: code = NotFound desc = could not find container \"a75e75394bad505488d6559266c3259e1e947ffca421dc069ad241dde82a7f95\": container with ID starting with a75e75394bad505488d6559266c3259e1e947ffca421dc069ad241dde82a7f95 not found: ID does not exist" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.145576 4559 scope.go:117] "RemoveContainer" containerID="46b8f2b2ff14784cbf9d6c17a44eb00e0191943703d04d3d075c37adec30ec5d" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.148887 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46b8f2b2ff14784cbf9d6c17a44eb00e0191943703d04d3d075c37adec30ec5d"} err="failed to get container status \"46b8f2b2ff14784cbf9d6c17a44eb00e0191943703d04d3d075c37adec30ec5d\": rpc error: code = NotFound desc = could not find container \"46b8f2b2ff14784cbf9d6c17a44eb00e0191943703d04d3d075c37adec30ec5d\": container with ID starting with 46b8f2b2ff14784cbf9d6c17a44eb00e0191943703d04d3d075c37adec30ec5d not found: ID does not exist" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.148919 4559 scope.go:117] "RemoveContainer" containerID="a75e75394bad505488d6559266c3259e1e947ffca421dc069ad241dde82a7f95" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.151103 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a75e75394bad505488d6559266c3259e1e947ffca421dc069ad241dde82a7f95"} err="failed to get container status \"a75e75394bad505488d6559266c3259e1e947ffca421dc069ad241dde82a7f95\": rpc error: code = NotFound desc = could not find container \"a75e75394bad505488d6559266c3259e1e947ffca421dc069ad241dde82a7f95\": container with ID starting with a75e75394bad505488d6559266c3259e1e947ffca421dc069ad241dde82a7f95 not found: ID does not exist" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.189167 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b8ac924-b4b7-4374-8f5f-7b2133936385-logs\") pod \"cinder-api-0\" (UID: \"3b8ac924-b4b7-4374-8f5f-7b2133936385\") " pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.189219 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8ac924-b4b7-4374-8f5f-7b2133936385-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3b8ac924-b4b7-4374-8f5f-7b2133936385\") " pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.189314 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b8ac924-b4b7-4374-8f5f-7b2133936385-config-data\") pod \"cinder-api-0\" (UID: \"3b8ac924-b4b7-4374-8f5f-7b2133936385\") " pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.189401 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b8ac924-b4b7-4374-8f5f-7b2133936385-public-tls-certs\") pod \"cinder-api-0\" (UID: \"3b8ac924-b4b7-4374-8f5f-7b2133936385\") " pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.189435 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3b8ac924-b4b7-4374-8f5f-7b2133936385-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3b8ac924-b4b7-4374-8f5f-7b2133936385\") " pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.189476 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw789\" (UniqueName: \"kubernetes.io/projected/3b8ac924-b4b7-4374-8f5f-7b2133936385-kube-api-access-dw789\") pod \"cinder-api-0\" (UID: \"3b8ac924-b4b7-4374-8f5f-7b2133936385\") " pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.189489 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b8ac924-b4b7-4374-8f5f-7b2133936385-config-data-custom\") pod \"cinder-api-0\" (UID: \"3b8ac924-b4b7-4374-8f5f-7b2133936385\") " pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.189541 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b8ac924-b4b7-4374-8f5f-7b2133936385-scripts\") pod \"cinder-api-0\" (UID: \"3b8ac924-b4b7-4374-8f5f-7b2133936385\") " pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.189577 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b8ac924-b4b7-4374-8f5f-7b2133936385-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"3b8ac924-b4b7-4374-8f5f-7b2133936385\") " pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.291481 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b8ac924-b4b7-4374-8f5f-7b2133936385-public-tls-certs\") pod \"cinder-api-0\" (UID: \"3b8ac924-b4b7-4374-8f5f-7b2133936385\") " pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.291528 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3b8ac924-b4b7-4374-8f5f-7b2133936385-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3b8ac924-b4b7-4374-8f5f-7b2133936385\") " pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.291567 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw789\" (UniqueName: \"kubernetes.io/projected/3b8ac924-b4b7-4374-8f5f-7b2133936385-kube-api-access-dw789\") pod \"cinder-api-0\" (UID: \"3b8ac924-b4b7-4374-8f5f-7b2133936385\") " pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.291584 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b8ac924-b4b7-4374-8f5f-7b2133936385-config-data-custom\") pod \"cinder-api-0\" (UID: \"3b8ac924-b4b7-4374-8f5f-7b2133936385\") " pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.291626 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b8ac924-b4b7-4374-8f5f-7b2133936385-scripts\") pod \"cinder-api-0\" (UID: \"3b8ac924-b4b7-4374-8f5f-7b2133936385\") " pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.291669 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b8ac924-b4b7-4374-8f5f-7b2133936385-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"3b8ac924-b4b7-4374-8f5f-7b2133936385\") " pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.291730 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b8ac924-b4b7-4374-8f5f-7b2133936385-logs\") pod \"cinder-api-0\" (UID: \"3b8ac924-b4b7-4374-8f5f-7b2133936385\") " pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.291752 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8ac924-b4b7-4374-8f5f-7b2133936385-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3b8ac924-b4b7-4374-8f5f-7b2133936385\") " pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.291826 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b8ac924-b4b7-4374-8f5f-7b2133936385-config-data\") pod \"cinder-api-0\" (UID: \"3b8ac924-b4b7-4374-8f5f-7b2133936385\") " pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.292258 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3b8ac924-b4b7-4374-8f5f-7b2133936385-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3b8ac924-b4b7-4374-8f5f-7b2133936385\") " pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.292418 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b8ac924-b4b7-4374-8f5f-7b2133936385-logs\") pod \"cinder-api-0\" (UID: \"3b8ac924-b4b7-4374-8f5f-7b2133936385\") " pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.296785 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b8ac924-b4b7-4374-8f5f-7b2133936385-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"3b8ac924-b4b7-4374-8f5f-7b2133936385\") " pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.297260 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b8ac924-b4b7-4374-8f5f-7b2133936385-public-tls-certs\") pod \"cinder-api-0\" (UID: \"3b8ac924-b4b7-4374-8f5f-7b2133936385\") " pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.297958 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b8ac924-b4b7-4374-8f5f-7b2133936385-scripts\") pod \"cinder-api-0\" (UID: \"3b8ac924-b4b7-4374-8f5f-7b2133936385\") " pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.298233 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b8ac924-b4b7-4374-8f5f-7b2133936385-config-data\") pod \"cinder-api-0\" (UID: \"3b8ac924-b4b7-4374-8f5f-7b2133936385\") " pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.298728 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b8ac924-b4b7-4374-8f5f-7b2133936385-config-data-custom\") pod \"cinder-api-0\" (UID: \"3b8ac924-b4b7-4374-8f5f-7b2133936385\") " pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.299146 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8ac924-b4b7-4374-8f5f-7b2133936385-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3b8ac924-b4b7-4374-8f5f-7b2133936385\") " pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.308592 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw789\" (UniqueName: \"kubernetes.io/projected/3b8ac924-b4b7-4374-8f5f-7b2133936385-kube-api-access-dw789\") pod \"cinder-api-0\" (UID: \"3b8ac924-b4b7-4374-8f5f-7b2133936385\") " pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.395882 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-nvnnf" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.421357 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.476088 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-55bb64d698-6lpkd" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.494165 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b5ec40f-303a-4c84-bf2a-8bf8991974df-config-volume\") pod \"5b5ec40f-303a-4c84-bf2a-8bf8991974df\" (UID: \"5b5ec40f-303a-4c84-bf2a-8bf8991974df\") " Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.494353 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b5ec40f-303a-4c84-bf2a-8bf8991974df-secret-volume\") pod \"5b5ec40f-303a-4c84-bf2a-8bf8991974df\" (UID: \"5b5ec40f-303a-4c84-bf2a-8bf8991974df\") " Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.494468 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plhvv\" (UniqueName: \"kubernetes.io/projected/5b5ec40f-303a-4c84-bf2a-8bf8991974df-kube-api-access-plhvv\") pod \"5b5ec40f-303a-4c84-bf2a-8bf8991974df\" (UID: \"5b5ec40f-303a-4c84-bf2a-8bf8991974df\") " Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.496940 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b5ec40f-303a-4c84-bf2a-8bf8991974df-config-volume" (OuterVolumeSpecName: "config-volume") pod "5b5ec40f-303a-4c84-bf2a-8bf8991974df" (UID: "5b5ec40f-303a-4c84-bf2a-8bf8991974df"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.498940 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b5ec40f-303a-4c84-bf2a-8bf8991974df-kube-api-access-plhvv" (OuterVolumeSpecName: "kube-api-access-plhvv") pod "5b5ec40f-303a-4c84-bf2a-8bf8991974df" (UID: "5b5ec40f-303a-4c84-bf2a-8bf8991974df"). InnerVolumeSpecName "kube-api-access-plhvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.500223 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b5ec40f-303a-4c84-bf2a-8bf8991974df-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5b5ec40f-303a-4c84-bf2a-8bf8991974df" (UID: "5b5ec40f-303a-4c84-bf2a-8bf8991974df"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.597707 4559 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b5ec40f-303a-4c84-bf2a-8bf8991974df-config-volume\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.597934 4559 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b5ec40f-303a-4c84-bf2a-8bf8991974df-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.597946 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plhvv\" (UniqueName: \"kubernetes.io/projected/5b5ec40f-303a-4c84-bf2a-8bf8991974df-kube-api-access-plhvv\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.672675 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-55bb64d698-6lpkd" Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.726173 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-798c5546f8-28xxs"] Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.726383 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-798c5546f8-28xxs" podUID="69cec8df-ef13-4a2a-884b-78ba11c34b3e" containerName="barbican-api-log" containerID="cri-o://b9714d8b8006b6c8af8b3d0c7c2cf22f1e859448a454ee2e74ac2bf1b2c6fcee" gracePeriod=30 Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.726452 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-798c5546f8-28xxs" podUID="69cec8df-ef13-4a2a-884b-78ba11c34b3e" containerName="barbican-api" containerID="cri-o://1c44ab5eb8f279614cdf3f57ebe976f6d3d88559796aa8cd24bd1e1462530a00" gracePeriod=30 Nov 23 07:00:03 crc kubenswrapper[4559]: I1123 07:00:03.856969 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 23 07:00:03 crc kubenswrapper[4559]: W1123 07:00:03.863569 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b8ac924_b4b7_4374_8f5f_7b2133936385.slice/crio-a4349c65481c73a7723bb07855b9fbe91f50ba00ebaa6fc06b0cbf3613493b1e WatchSource:0}: Error finding container a4349c65481c73a7723bb07855b9fbe91f50ba00ebaa6fc06b0cbf3613493b1e: Status 404 returned error can't find the container with id a4349c65481c73a7723bb07855b9fbe91f50ba00ebaa6fc06b0cbf3613493b1e Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.050710 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-nvnnf" event={"ID":"5b5ec40f-303a-4c84-bf2a-8bf8991974df","Type":"ContainerDied","Data":"2355633c5f2024a7a843ff258274db073c1e237924af9c433b08f0b06b6d0dd9"} Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.050994 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2355633c5f2024a7a843ff258274db073c1e237924af9c433b08f0b06b6d0dd9" Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.050910 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398020-nvnnf" Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.051957 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3b8ac924-b4b7-4374-8f5f-7b2133936385","Type":"ContainerStarted","Data":"a4349c65481c73a7723bb07855b9fbe91f50ba00ebaa6fc06b0cbf3613493b1e"} Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.058172 4559 generic.go:334] "Generic (PLEG): container finished" podID="69cec8df-ef13-4a2a-884b-78ba11c34b3e" containerID="b9714d8b8006b6c8af8b3d0c7c2cf22f1e859448a454ee2e74ac2bf1b2c6fcee" exitCode=143 Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.058248 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-798c5546f8-28xxs" event={"ID":"69cec8df-ef13-4a2a-884b-78ba11c34b3e","Type":"ContainerDied","Data":"b9714d8b8006b6c8af8b3d0c7c2cf22f1e859448a454ee2e74ac2bf1b2c6fcee"} Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.061894 4559 generic.go:334] "Generic (PLEG): container finished" podID="4163bfe8-0de7-4490-b825-55d0f0343787" containerID="cd785a0026c59d5e74c27422561bc4c114f548239eeb6189968a2dd70cd1543d" exitCode=0 Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.061960 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4163bfe8-0de7-4490-b825-55d0f0343787","Type":"ContainerDied","Data":"cd785a0026c59d5e74c27422561bc4c114f548239eeb6189968a2dd70cd1543d"} Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.077747 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.112611 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4163bfe8-0de7-4490-b825-55d0f0343787-scripts\") pod \"4163bfe8-0de7-4490-b825-55d0f0343787\" (UID: \"4163bfe8-0de7-4490-b825-55d0f0343787\") " Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.112785 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4163bfe8-0de7-4490-b825-55d0f0343787-run-httpd\") pod \"4163bfe8-0de7-4490-b825-55d0f0343787\" (UID: \"4163bfe8-0de7-4490-b825-55d0f0343787\") " Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.112843 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4163bfe8-0de7-4490-b825-55d0f0343787-sg-core-conf-yaml\") pod \"4163bfe8-0de7-4490-b825-55d0f0343787\" (UID: \"4163bfe8-0de7-4490-b825-55d0f0343787\") " Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.112868 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf5bp\" (UniqueName: \"kubernetes.io/projected/4163bfe8-0de7-4490-b825-55d0f0343787-kube-api-access-gf5bp\") pod \"4163bfe8-0de7-4490-b825-55d0f0343787\" (UID: \"4163bfe8-0de7-4490-b825-55d0f0343787\") " Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.112958 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4163bfe8-0de7-4490-b825-55d0f0343787-config-data\") pod \"4163bfe8-0de7-4490-b825-55d0f0343787\" (UID: \"4163bfe8-0de7-4490-b825-55d0f0343787\") " Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.113015 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4163bfe8-0de7-4490-b825-55d0f0343787-combined-ca-bundle\") pod \"4163bfe8-0de7-4490-b825-55d0f0343787\" (UID: \"4163bfe8-0de7-4490-b825-55d0f0343787\") " Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.113042 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4163bfe8-0de7-4490-b825-55d0f0343787-log-httpd\") pod \"4163bfe8-0de7-4490-b825-55d0f0343787\" (UID: \"4163bfe8-0de7-4490-b825-55d0f0343787\") " Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.116376 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4163bfe8-0de7-4490-b825-55d0f0343787-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4163bfe8-0de7-4490-b825-55d0f0343787" (UID: "4163bfe8-0de7-4490-b825-55d0f0343787"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.118072 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4163bfe8-0de7-4490-b825-55d0f0343787-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4163bfe8-0de7-4490-b825-55d0f0343787" (UID: "4163bfe8-0de7-4490-b825-55d0f0343787"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.122498 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4163bfe8-0de7-4490-b825-55d0f0343787-scripts" (OuterVolumeSpecName: "scripts") pod "4163bfe8-0de7-4490-b825-55d0f0343787" (UID: "4163bfe8-0de7-4490-b825-55d0f0343787"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.122782 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4163bfe8-0de7-4490-b825-55d0f0343787-kube-api-access-gf5bp" (OuterVolumeSpecName: "kube-api-access-gf5bp") pod "4163bfe8-0de7-4490-b825-55d0f0343787" (UID: "4163bfe8-0de7-4490-b825-55d0f0343787"). InnerVolumeSpecName "kube-api-access-gf5bp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.139018 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4163bfe8-0de7-4490-b825-55d0f0343787-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4163bfe8-0de7-4490-b825-55d0f0343787" (UID: "4163bfe8-0de7-4490-b825-55d0f0343787"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.176605 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4163bfe8-0de7-4490-b825-55d0f0343787-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4163bfe8-0de7-4490-b825-55d0f0343787" (UID: "4163bfe8-0de7-4490-b825-55d0f0343787"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.192387 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4163bfe8-0de7-4490-b825-55d0f0343787-config-data" (OuterVolumeSpecName: "config-data") pod "4163bfe8-0de7-4490-b825-55d0f0343787" (UID: "4163bfe8-0de7-4490-b825-55d0f0343787"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.216045 4559 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4163bfe8-0de7-4490-b825-55d0f0343787-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.216191 4559 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4163bfe8-0de7-4490-b825-55d0f0343787-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.216202 4559 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4163bfe8-0de7-4490-b825-55d0f0343787-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.216211 4559 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4163bfe8-0de7-4490-b825-55d0f0343787-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.216221 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf5bp\" (UniqueName: \"kubernetes.io/projected/4163bfe8-0de7-4490-b825-55d0f0343787-kube-api-access-gf5bp\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.216249 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4163bfe8-0de7-4490-b825-55d0f0343787-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.216260 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4163bfe8-0de7-4490-b825-55d0f0343787-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.284631 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aec4f5ef-bdf5-475a-8a2b-5d70751161c6" path="/var/lib/kubelet/pods/aec4f5ef-bdf5-475a-8a2b-5d70751161c6/volumes" Nov 23 07:00:04 crc kubenswrapper[4559]: I1123 07:00:04.547101 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.073010 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3b8ac924-b4b7-4374-8f5f-7b2133936385","Type":"ContainerStarted","Data":"306b3fe5e74b2301974660e58b4e2c943cf24ff12c3e02434a4870ff1d6e0f56"} Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.073071 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3b8ac924-b4b7-4374-8f5f-7b2133936385","Type":"ContainerStarted","Data":"466004e266a1b73ab963a7ac64d0cc9dca86cadceceef21fb67992b1508ab74c"} Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.073170 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.077357 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4163bfe8-0de7-4490-b825-55d0f0343787","Type":"ContainerDied","Data":"abd76a57cf9f3e9befff8730283bb9ea4c4f010c2d5207130c0f760b6cd1fa84"} Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.077380 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.077417 4559 scope.go:117] "RemoveContainer" containerID="45189c2b7acb1f8b18c8954aedcd433313f126fd67dde19896cb9e5e255dee02" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.115530 4559 scope.go:117] "RemoveContainer" containerID="5fb7c20c88f83d97306e5172d980cc901c730345621912965ed251efeb54307a" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.142428 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.142406087 podStartE2EDuration="2.142406087s" podCreationTimestamp="2025-11-23 07:00:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:00:05.106941908 +0000 UTC m=+907.128927522" watchObservedRunningTime="2025-11-23 07:00:05.142406087 +0000 UTC m=+907.164391701" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.147954 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.159698 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.161953 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:00:05 crc kubenswrapper[4559]: E1123 07:00:05.163886 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4163bfe8-0de7-4490-b825-55d0f0343787" containerName="ceilometer-central-agent" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.163981 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="4163bfe8-0de7-4490-b825-55d0f0343787" containerName="ceilometer-central-agent" Nov 23 07:00:05 crc kubenswrapper[4559]: E1123 07:00:05.164084 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4163bfe8-0de7-4490-b825-55d0f0343787" containerName="proxy-httpd" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.164151 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="4163bfe8-0de7-4490-b825-55d0f0343787" containerName="proxy-httpd" Nov 23 07:00:05 crc kubenswrapper[4559]: E1123 07:00:05.164744 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4163bfe8-0de7-4490-b825-55d0f0343787" containerName="sg-core" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.164838 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="4163bfe8-0de7-4490-b825-55d0f0343787" containerName="sg-core" Nov 23 07:00:05 crc kubenswrapper[4559]: E1123 07:00:05.164933 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4163bfe8-0de7-4490-b825-55d0f0343787" containerName="ceilometer-notification-agent" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.164999 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="4163bfe8-0de7-4490-b825-55d0f0343787" containerName="ceilometer-notification-agent" Nov 23 07:00:05 crc kubenswrapper[4559]: E1123 07:00:05.165054 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b5ec40f-303a-4c84-bf2a-8bf8991974df" containerName="collect-profiles" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.165117 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b5ec40f-303a-4c84-bf2a-8bf8991974df" containerName="collect-profiles" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.165765 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="4163bfe8-0de7-4490-b825-55d0f0343787" containerName="sg-core" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.165864 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="4163bfe8-0de7-4490-b825-55d0f0343787" containerName="proxy-httpd" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.165960 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b5ec40f-303a-4c84-bf2a-8bf8991974df" containerName="collect-profiles" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.166041 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="4163bfe8-0de7-4490-b825-55d0f0343787" containerName="ceilometer-notification-agent" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.166118 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="4163bfe8-0de7-4490-b825-55d0f0343787" containerName="ceilometer-central-agent" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.168460 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.173160 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.174208 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.176509 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.192806 4559 scope.go:117] "RemoveContainer" containerID="cd785a0026c59d5e74c27422561bc4c114f548239eeb6189968a2dd70cd1543d" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.214301 4559 scope.go:117] "RemoveContainer" containerID="ba10afde1644658d448c1bb320ea05c40fdfb8995a31b98a833593e1267bbfbe" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.247693 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-824mc\" (UniqueName: \"kubernetes.io/projected/5b0ee294-003c-49b1-878e-50097454795f-kube-api-access-824mc\") pod \"ceilometer-0\" (UID: \"5b0ee294-003c-49b1-878e-50097454795f\") " pod="openstack/ceilometer-0" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.247823 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b0ee294-003c-49b1-878e-50097454795f-scripts\") pod \"ceilometer-0\" (UID: \"5b0ee294-003c-49b1-878e-50097454795f\") " pod="openstack/ceilometer-0" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.247861 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b0ee294-003c-49b1-878e-50097454795f-log-httpd\") pod \"ceilometer-0\" (UID: \"5b0ee294-003c-49b1-878e-50097454795f\") " pod="openstack/ceilometer-0" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.248061 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b0ee294-003c-49b1-878e-50097454795f-config-data\") pod \"ceilometer-0\" (UID: \"5b0ee294-003c-49b1-878e-50097454795f\") " pod="openstack/ceilometer-0" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.248231 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b0ee294-003c-49b1-878e-50097454795f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5b0ee294-003c-49b1-878e-50097454795f\") " pod="openstack/ceilometer-0" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.248331 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b0ee294-003c-49b1-878e-50097454795f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5b0ee294-003c-49b1-878e-50097454795f\") " pod="openstack/ceilometer-0" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.248418 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b0ee294-003c-49b1-878e-50097454795f-run-httpd\") pod \"ceilometer-0\" (UID: \"5b0ee294-003c-49b1-878e-50097454795f\") " pod="openstack/ceilometer-0" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.349819 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-824mc\" (UniqueName: \"kubernetes.io/projected/5b0ee294-003c-49b1-878e-50097454795f-kube-api-access-824mc\") pod \"ceilometer-0\" (UID: \"5b0ee294-003c-49b1-878e-50097454795f\") " pod="openstack/ceilometer-0" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.349872 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b0ee294-003c-49b1-878e-50097454795f-log-httpd\") pod \"ceilometer-0\" (UID: \"5b0ee294-003c-49b1-878e-50097454795f\") " pod="openstack/ceilometer-0" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.349900 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b0ee294-003c-49b1-878e-50097454795f-scripts\") pod \"ceilometer-0\" (UID: \"5b0ee294-003c-49b1-878e-50097454795f\") " pod="openstack/ceilometer-0" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.349954 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b0ee294-003c-49b1-878e-50097454795f-config-data\") pod \"ceilometer-0\" (UID: \"5b0ee294-003c-49b1-878e-50097454795f\") " pod="openstack/ceilometer-0" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.350008 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b0ee294-003c-49b1-878e-50097454795f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5b0ee294-003c-49b1-878e-50097454795f\") " pod="openstack/ceilometer-0" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.350043 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b0ee294-003c-49b1-878e-50097454795f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5b0ee294-003c-49b1-878e-50097454795f\") " pod="openstack/ceilometer-0" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.350077 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b0ee294-003c-49b1-878e-50097454795f-run-httpd\") pod \"ceilometer-0\" (UID: \"5b0ee294-003c-49b1-878e-50097454795f\") " pod="openstack/ceilometer-0" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.350513 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b0ee294-003c-49b1-878e-50097454795f-run-httpd\") pod \"ceilometer-0\" (UID: \"5b0ee294-003c-49b1-878e-50097454795f\") " pod="openstack/ceilometer-0" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.351284 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b0ee294-003c-49b1-878e-50097454795f-log-httpd\") pod \"ceilometer-0\" (UID: \"5b0ee294-003c-49b1-878e-50097454795f\") " pod="openstack/ceilometer-0" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.355774 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b0ee294-003c-49b1-878e-50097454795f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5b0ee294-003c-49b1-878e-50097454795f\") " pod="openstack/ceilometer-0" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.356012 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b0ee294-003c-49b1-878e-50097454795f-config-data\") pod \"ceilometer-0\" (UID: \"5b0ee294-003c-49b1-878e-50097454795f\") " pod="openstack/ceilometer-0" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.356565 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b0ee294-003c-49b1-878e-50097454795f-scripts\") pod \"ceilometer-0\" (UID: \"5b0ee294-003c-49b1-878e-50097454795f\") " pod="openstack/ceilometer-0" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.356961 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b0ee294-003c-49b1-878e-50097454795f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5b0ee294-003c-49b1-878e-50097454795f\") " pod="openstack/ceilometer-0" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.364347 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-824mc\" (UniqueName: \"kubernetes.io/projected/5b0ee294-003c-49b1-878e-50097454795f-kube-api-access-824mc\") pod \"ceilometer-0\" (UID: \"5b0ee294-003c-49b1-878e-50097454795f\") " pod="openstack/ceilometer-0" Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.504444 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:00:05 crc kubenswrapper[4559]: W1123 07:00:05.869518 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b0ee294_003c_49b1_878e_50097454795f.slice/crio-9381f2fde154e6c805142313edda953a39caa0a3264e48406db0f56b6604ad69 WatchSource:0}: Error finding container 9381f2fde154e6c805142313edda953a39caa0a3264e48406db0f56b6604ad69: Status 404 returned error can't find the container with id 9381f2fde154e6c805142313edda953a39caa0a3264e48406db0f56b6604ad69 Nov 23 07:00:05 crc kubenswrapper[4559]: I1123 07:00:05.869536 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:00:06 crc kubenswrapper[4559]: I1123 07:00:06.088177 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b0ee294-003c-49b1-878e-50097454795f","Type":"ContainerStarted","Data":"9381f2fde154e6c805142313edda953a39caa0a3264e48406db0f56b6604ad69"} Nov 23 07:00:06 crc kubenswrapper[4559]: I1123 07:00:06.285634 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4163bfe8-0de7-4490-b825-55d0f0343787" path="/var/lib/kubelet/pods/4163bfe8-0de7-4490-b825-55d0f0343787/volumes" Nov 23 07:00:06 crc kubenswrapper[4559]: I1123 07:00:06.860829 4559 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-798c5546f8-28xxs" podUID="69cec8df-ef13-4a2a-884b-78ba11c34b3e" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.154:9311/healthcheck\": read tcp 10.217.0.2:42890->10.217.0.154:9311: read: connection reset by peer" Nov 23 07:00:06 crc kubenswrapper[4559]: I1123 07:00:06.860850 4559 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-798c5546f8-28xxs" podUID="69cec8df-ef13-4a2a-884b-78ba11c34b3e" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.154:9311/healthcheck\": read tcp 10.217.0.2:42876->10.217.0.154:9311: read: connection reset by peer" Nov 23 07:00:06 crc kubenswrapper[4559]: I1123 07:00:06.923280 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-67f8477cdd-4x7mh" Nov 23 07:00:07 crc kubenswrapper[4559]: I1123 07:00:07.096877 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b0ee294-003c-49b1-878e-50097454795f","Type":"ContainerStarted","Data":"c21b860c45243f8824cd1b77c0a54694e67b445eac3577a219d0e5d4f7e8ce15"} Nov 23 07:00:07 crc kubenswrapper[4559]: I1123 07:00:07.105301 4559 generic.go:334] "Generic (PLEG): container finished" podID="69cec8df-ef13-4a2a-884b-78ba11c34b3e" containerID="1c44ab5eb8f279614cdf3f57ebe976f6d3d88559796aa8cd24bd1e1462530a00" exitCode=0 Nov 23 07:00:07 crc kubenswrapper[4559]: I1123 07:00:07.105339 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-798c5546f8-28xxs" event={"ID":"69cec8df-ef13-4a2a-884b-78ba11c34b3e","Type":"ContainerDied","Data":"1c44ab5eb8f279614cdf3f57ebe976f6d3d88559796aa8cd24bd1e1462530a00"} Nov 23 07:00:07 crc kubenswrapper[4559]: I1123 07:00:07.281147 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-798c5546f8-28xxs" Nov 23 07:00:07 crc kubenswrapper[4559]: I1123 07:00:07.391368 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69cec8df-ef13-4a2a-884b-78ba11c34b3e-config-data\") pod \"69cec8df-ef13-4a2a-884b-78ba11c34b3e\" (UID: \"69cec8df-ef13-4a2a-884b-78ba11c34b3e\") " Nov 23 07:00:07 crc kubenswrapper[4559]: I1123 07:00:07.391471 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bttvb\" (UniqueName: \"kubernetes.io/projected/69cec8df-ef13-4a2a-884b-78ba11c34b3e-kube-api-access-bttvb\") pod \"69cec8df-ef13-4a2a-884b-78ba11c34b3e\" (UID: \"69cec8df-ef13-4a2a-884b-78ba11c34b3e\") " Nov 23 07:00:07 crc kubenswrapper[4559]: I1123 07:00:07.391610 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69cec8df-ef13-4a2a-884b-78ba11c34b3e-logs\") pod \"69cec8df-ef13-4a2a-884b-78ba11c34b3e\" (UID: \"69cec8df-ef13-4a2a-884b-78ba11c34b3e\") " Nov 23 07:00:07 crc kubenswrapper[4559]: I1123 07:00:07.391681 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69cec8df-ef13-4a2a-884b-78ba11c34b3e-combined-ca-bundle\") pod \"69cec8df-ef13-4a2a-884b-78ba11c34b3e\" (UID: \"69cec8df-ef13-4a2a-884b-78ba11c34b3e\") " Nov 23 07:00:07 crc kubenswrapper[4559]: I1123 07:00:07.391704 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69cec8df-ef13-4a2a-884b-78ba11c34b3e-config-data-custom\") pod \"69cec8df-ef13-4a2a-884b-78ba11c34b3e\" (UID: \"69cec8df-ef13-4a2a-884b-78ba11c34b3e\") " Nov 23 07:00:07 crc kubenswrapper[4559]: I1123 07:00:07.393313 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69cec8df-ef13-4a2a-884b-78ba11c34b3e-logs" (OuterVolumeSpecName: "logs") pod "69cec8df-ef13-4a2a-884b-78ba11c34b3e" (UID: "69cec8df-ef13-4a2a-884b-78ba11c34b3e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:00:07 crc kubenswrapper[4559]: I1123 07:00:07.396034 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69cec8df-ef13-4a2a-884b-78ba11c34b3e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "69cec8df-ef13-4a2a-884b-78ba11c34b3e" (UID: "69cec8df-ef13-4a2a-884b-78ba11c34b3e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:07 crc kubenswrapper[4559]: I1123 07:00:07.396512 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69cec8df-ef13-4a2a-884b-78ba11c34b3e-kube-api-access-bttvb" (OuterVolumeSpecName: "kube-api-access-bttvb") pod "69cec8df-ef13-4a2a-884b-78ba11c34b3e" (UID: "69cec8df-ef13-4a2a-884b-78ba11c34b3e"). InnerVolumeSpecName "kube-api-access-bttvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:00:07 crc kubenswrapper[4559]: I1123 07:00:07.440121 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69cec8df-ef13-4a2a-884b-78ba11c34b3e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "69cec8df-ef13-4a2a-884b-78ba11c34b3e" (UID: "69cec8df-ef13-4a2a-884b-78ba11c34b3e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:07 crc kubenswrapper[4559]: I1123 07:00:07.462007 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69cec8df-ef13-4a2a-884b-78ba11c34b3e-config-data" (OuterVolumeSpecName: "config-data") pod "69cec8df-ef13-4a2a-884b-78ba11c34b3e" (UID: "69cec8df-ef13-4a2a-884b-78ba11c34b3e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:07 crc kubenswrapper[4559]: I1123 07:00:07.493886 4559 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69cec8df-ef13-4a2a-884b-78ba11c34b3e-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:07 crc kubenswrapper[4559]: I1123 07:00:07.493909 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69cec8df-ef13-4a2a-884b-78ba11c34b3e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:07 crc kubenswrapper[4559]: I1123 07:00:07.493919 4559 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69cec8df-ef13-4a2a-884b-78ba11c34b3e-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:07 crc kubenswrapper[4559]: I1123 07:00:07.493936 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69cec8df-ef13-4a2a-884b-78ba11c34b3e-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:07 crc kubenswrapper[4559]: I1123 07:00:07.493945 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bttvb\" (UniqueName: \"kubernetes.io/projected/69cec8df-ef13-4a2a-884b-78ba11c34b3e-kube-api-access-bttvb\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:08 crc kubenswrapper[4559]: I1123 07:00:08.118894 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b0ee294-003c-49b1-878e-50097454795f","Type":"ContainerStarted","Data":"46666f0629352c2d81b02e541670d6640a217e0d5b21a838ed83703abd0d91ef"} Nov 23 07:00:08 crc kubenswrapper[4559]: I1123 07:00:08.122052 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-798c5546f8-28xxs" event={"ID":"69cec8df-ef13-4a2a-884b-78ba11c34b3e","Type":"ContainerDied","Data":"0a93f15632ff4ff14ff7816dae4e4f21250c232e2843dc579ae2530b9144b12b"} Nov 23 07:00:08 crc kubenswrapper[4559]: I1123 07:00:08.122097 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-798c5546f8-28xxs" Nov 23 07:00:08 crc kubenswrapper[4559]: I1123 07:00:08.122105 4559 scope.go:117] "RemoveContainer" containerID="1c44ab5eb8f279614cdf3f57ebe976f6d3d88559796aa8cd24bd1e1462530a00" Nov 23 07:00:08 crc kubenswrapper[4559]: I1123 07:00:08.141812 4559 scope.go:117] "RemoveContainer" containerID="b9714d8b8006b6c8af8b3d0c7c2cf22f1e859448a454ee2e74ac2bf1b2c6fcee" Nov 23 07:00:08 crc kubenswrapper[4559]: I1123 07:00:08.158433 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-798c5546f8-28xxs"] Nov 23 07:00:08 crc kubenswrapper[4559]: I1123 07:00:08.164311 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-798c5546f8-28xxs"] Nov 23 07:00:08 crc kubenswrapper[4559]: I1123 07:00:08.309148 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69cec8df-ef13-4a2a-884b-78ba11c34b3e" path="/var/lib/kubelet/pods/69cec8df-ef13-4a2a-884b-78ba11c34b3e/volumes" Nov 23 07:00:08 crc kubenswrapper[4559]: I1123 07:00:08.893355 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-d8c98f585-kr7vt" Nov 23 07:00:08 crc kubenswrapper[4559]: I1123 07:00:08.950270 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-67f8477cdd-4x7mh"] Nov 23 07:00:08 crc kubenswrapper[4559]: I1123 07:00:08.950563 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-67f8477cdd-4x7mh" podUID="b649ece4-50a8-4449-ba98-2815000866fa" containerName="neutron-api" containerID="cri-o://bd5950999e0816dc3a6c6c59ae79a87d4896d498e6aec37ee3fb20cea46e18c3" gracePeriod=30 Nov 23 07:00:08 crc kubenswrapper[4559]: I1123 07:00:08.951088 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-67f8477cdd-4x7mh" podUID="b649ece4-50a8-4449-ba98-2815000866fa" containerName="neutron-httpd" containerID="cri-o://70a6af06589557c859bf0b9b1efe59d4f712ca7df817304e61ba10633adc6a9b" gracePeriod=30 Nov 23 07:00:09 crc kubenswrapper[4559]: I1123 07:00:09.133922 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b0ee294-003c-49b1-878e-50097454795f","Type":"ContainerStarted","Data":"ae1b972ec5cc95c188cb0fac2c7e0ec994b95f2fa2655a84967d824390d1ceb4"} Nov 23 07:00:09 crc kubenswrapper[4559]: I1123 07:00:09.609805 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" Nov 23 07:00:09 crc kubenswrapper[4559]: I1123 07:00:09.658548 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ccdb5d4d7-59qkv"] Nov 23 07:00:09 crc kubenswrapper[4559]: I1123 07:00:09.658827 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" podUID="5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0" containerName="dnsmasq-dns" containerID="cri-o://d12b26a69c747e43b76d9f3e19c15607447fb89167840665827bd36455f11a6a" gracePeriod=10 Nov 23 07:00:09 crc kubenswrapper[4559]: I1123 07:00:09.746479 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 23 07:00:09 crc kubenswrapper[4559]: I1123 07:00:09.826692 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.143672 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.144922 4559 generic.go:334] "Generic (PLEG): container finished" podID="5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0" containerID="d12b26a69c747e43b76d9f3e19c15607447fb89167840665827bd36455f11a6a" exitCode=0 Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.145007 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" event={"ID":"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0","Type":"ContainerDied","Data":"d12b26a69c747e43b76d9f3e19c15607447fb89167840665827bd36455f11a6a"} Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.145075 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" event={"ID":"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0","Type":"ContainerDied","Data":"6d4c8038b911c2531b576d54af5fae2b9791edcbda45bf79948e6f8dd712af35"} Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.145108 4559 scope.go:117] "RemoveContainer" containerID="d12b26a69c747e43b76d9f3e19c15607447fb89167840665827bd36455f11a6a" Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.147819 4559 generic.go:334] "Generic (PLEG): container finished" podID="b649ece4-50a8-4449-ba98-2815000866fa" containerID="70a6af06589557c859bf0b9b1efe59d4f712ca7df817304e61ba10633adc6a9b" exitCode=0 Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.147874 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67f8477cdd-4x7mh" event={"ID":"b649ece4-50a8-4449-ba98-2815000866fa","Type":"ContainerDied","Data":"70a6af06589557c859bf0b9b1efe59d4f712ca7df817304e61ba10633adc6a9b"} Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.151585 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="cb5e8b5b-0513-4392-a079-0dab7664ec84" containerName="cinder-scheduler" containerID="cri-o://1d0ef4f29080e8dba7e9ab6b9fc27101b075c4bc0291cf4f4b8f734983e8543a" gracePeriod=30 Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.152115 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b0ee294-003c-49b1-878e-50097454795f","Type":"ContainerStarted","Data":"81ccb4b66519f8f7d2facac1d0226906cf29314b35e5f091f2c049fa273ef2e2"} Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.152303 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.152154 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="cb5e8b5b-0513-4392-a079-0dab7664ec84" containerName="probe" containerID="cri-o://704862b0ff2fe684d984c817f5989629201f75ed385a420513da5c764edd5a15" gracePeriod=30 Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.177418 4559 scope.go:117] "RemoveContainer" containerID="58991aecfc4b746c4c67f4eda4d6dd9ee6884882102e6c8e63641970744f2578" Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.190176 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.866203761 podStartE2EDuration="5.190159636s" podCreationTimestamp="2025-11-23 07:00:05 +0000 UTC" firstStartedPulling="2025-11-23 07:00:05.871946258 +0000 UTC m=+907.893931873" lastFinishedPulling="2025-11-23 07:00:09.195902144 +0000 UTC m=+911.217887748" observedRunningTime="2025-11-23 07:00:10.181677543 +0000 UTC m=+912.203663157" watchObservedRunningTime="2025-11-23 07:00:10.190159636 +0000 UTC m=+912.212145249" Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.206957 4559 scope.go:117] "RemoveContainer" containerID="d12b26a69c747e43b76d9f3e19c15607447fb89167840665827bd36455f11a6a" Nov 23 07:00:10 crc kubenswrapper[4559]: E1123 07:00:10.208322 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d12b26a69c747e43b76d9f3e19c15607447fb89167840665827bd36455f11a6a\": container with ID starting with d12b26a69c747e43b76d9f3e19c15607447fb89167840665827bd36455f11a6a not found: ID does not exist" containerID="d12b26a69c747e43b76d9f3e19c15607447fb89167840665827bd36455f11a6a" Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.208365 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d12b26a69c747e43b76d9f3e19c15607447fb89167840665827bd36455f11a6a"} err="failed to get container status \"d12b26a69c747e43b76d9f3e19c15607447fb89167840665827bd36455f11a6a\": rpc error: code = NotFound desc = could not find container \"d12b26a69c747e43b76d9f3e19c15607447fb89167840665827bd36455f11a6a\": container with ID starting with d12b26a69c747e43b76d9f3e19c15607447fb89167840665827bd36455f11a6a not found: ID does not exist" Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.208385 4559 scope.go:117] "RemoveContainer" containerID="58991aecfc4b746c4c67f4eda4d6dd9ee6884882102e6c8e63641970744f2578" Nov 23 07:00:10 crc kubenswrapper[4559]: E1123 07:00:10.208708 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58991aecfc4b746c4c67f4eda4d6dd9ee6884882102e6c8e63641970744f2578\": container with ID starting with 58991aecfc4b746c4c67f4eda4d6dd9ee6884882102e6c8e63641970744f2578 not found: ID does not exist" containerID="58991aecfc4b746c4c67f4eda4d6dd9ee6884882102e6c8e63641970744f2578" Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.208748 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58991aecfc4b746c4c67f4eda4d6dd9ee6884882102e6c8e63641970744f2578"} err="failed to get container status \"58991aecfc4b746c4c67f4eda4d6dd9ee6884882102e6c8e63641970744f2578\": rpc error: code = NotFound desc = could not find container \"58991aecfc4b746c4c67f4eda4d6dd9ee6884882102e6c8e63641970744f2578\": container with ID starting with 58991aecfc4b746c4c67f4eda4d6dd9ee6884882102e6c8e63641970744f2578 not found: ID does not exist" Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.274878 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-dns-swift-storage-0\") pod \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\" (UID: \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\") " Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.274938 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-ovsdbserver-sb\") pod \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\" (UID: \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\") " Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.275017 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-dns-svc\") pod \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\" (UID: \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\") " Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.275139 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-ovsdbserver-nb\") pod \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\" (UID: \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\") " Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.275172 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p42gk\" (UniqueName: \"kubernetes.io/projected/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-kube-api-access-p42gk\") pod \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\" (UID: \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\") " Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.275227 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-config\") pod \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\" (UID: \"5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0\") " Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.282060 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-kube-api-access-p42gk" (OuterVolumeSpecName: "kube-api-access-p42gk") pod "5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0" (UID: "5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0"). InnerVolumeSpecName "kube-api-access-p42gk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.323401 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0" (UID: "5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.327500 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0" (UID: "5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.334388 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0" (UID: "5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.338988 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0" (UID: "5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.365147 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-config" (OuterVolumeSpecName: "config") pod "5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0" (UID: "5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.377303 4559 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.377331 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p42gk\" (UniqueName: \"kubernetes.io/projected/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-kube-api-access-p42gk\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.377470 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.377707 4559 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.377718 4559 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:10 crc kubenswrapper[4559]: I1123 07:00:10.377727 4559 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:11 crc kubenswrapper[4559]: I1123 07:00:11.162480 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ccdb5d4d7-59qkv" Nov 23 07:00:11 crc kubenswrapper[4559]: I1123 07:00:11.165247 4559 generic.go:334] "Generic (PLEG): container finished" podID="cb5e8b5b-0513-4392-a079-0dab7664ec84" containerID="704862b0ff2fe684d984c817f5989629201f75ed385a420513da5c764edd5a15" exitCode=0 Nov 23 07:00:11 crc kubenswrapper[4559]: I1123 07:00:11.165326 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb5e8b5b-0513-4392-a079-0dab7664ec84","Type":"ContainerDied","Data":"704862b0ff2fe684d984c817f5989629201f75ed385a420513da5c764edd5a15"} Nov 23 07:00:11 crc kubenswrapper[4559]: I1123 07:00:11.193632 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ccdb5d4d7-59qkv"] Nov 23 07:00:11 crc kubenswrapper[4559]: I1123 07:00:11.202192 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7ccdb5d4d7-59qkv"] Nov 23 07:00:11 crc kubenswrapper[4559]: I1123 07:00:11.772776 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7667f44776-mszxt" Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.183774 4559 generic.go:334] "Generic (PLEG): container finished" podID="b649ece4-50a8-4449-ba98-2815000866fa" containerID="bd5950999e0816dc3a6c6c59ae79a87d4896d498e6aec37ee3fb20cea46e18c3" exitCode=0 Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.183827 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67f8477cdd-4x7mh" event={"ID":"b649ece4-50a8-4449-ba98-2815000866fa","Type":"ContainerDied","Data":"bd5950999e0816dc3a6c6c59ae79a87d4896d498e6aec37ee3fb20cea46e18c3"} Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.285892 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0" path="/var/lib/kubelet/pods/5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0/volumes" Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.623853 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-67f8477cdd-4x7mh" Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.722393 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b649ece4-50a8-4449-ba98-2815000866fa-httpd-config\") pod \"b649ece4-50a8-4449-ba98-2815000866fa\" (UID: \"b649ece4-50a8-4449-ba98-2815000866fa\") " Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.722445 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b649ece4-50a8-4449-ba98-2815000866fa-config\") pod \"b649ece4-50a8-4449-ba98-2815000866fa\" (UID: \"b649ece4-50a8-4449-ba98-2815000866fa\") " Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.722476 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4f62\" (UniqueName: \"kubernetes.io/projected/b649ece4-50a8-4449-ba98-2815000866fa-kube-api-access-t4f62\") pod \"b649ece4-50a8-4449-ba98-2815000866fa\" (UID: \"b649ece4-50a8-4449-ba98-2815000866fa\") " Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.722515 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b649ece4-50a8-4449-ba98-2815000866fa-ovndb-tls-certs\") pod \"b649ece4-50a8-4449-ba98-2815000866fa\" (UID: \"b649ece4-50a8-4449-ba98-2815000866fa\") " Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.722747 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b649ece4-50a8-4449-ba98-2815000866fa-combined-ca-bundle\") pod \"b649ece4-50a8-4449-ba98-2815000866fa\" (UID: \"b649ece4-50a8-4449-ba98-2815000866fa\") " Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.723893 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.731869 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b649ece4-50a8-4449-ba98-2815000866fa-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "b649ece4-50a8-4449-ba98-2815000866fa" (UID: "b649ece4-50a8-4449-ba98-2815000866fa"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.731901 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b649ece4-50a8-4449-ba98-2815000866fa-kube-api-access-t4f62" (OuterVolumeSpecName: "kube-api-access-t4f62") pod "b649ece4-50a8-4449-ba98-2815000866fa" (UID: "b649ece4-50a8-4449-ba98-2815000866fa"). InnerVolumeSpecName "kube-api-access-t4f62". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.772180 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b649ece4-50a8-4449-ba98-2815000866fa-config" (OuterVolumeSpecName: "config") pod "b649ece4-50a8-4449-ba98-2815000866fa" (UID: "b649ece4-50a8-4449-ba98-2815000866fa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.781228 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b649ece4-50a8-4449-ba98-2815000866fa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b649ece4-50a8-4449-ba98-2815000866fa" (UID: "b649ece4-50a8-4449-ba98-2815000866fa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.789418 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7667f44776-mszxt" Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.792271 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b649ece4-50a8-4449-ba98-2815000866fa-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "b649ece4-50a8-4449-ba98-2815000866fa" (UID: "b649ece4-50a8-4449-ba98-2815000866fa"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.824050 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb5e8b5b-0513-4392-a079-0dab7664ec84-scripts\") pod \"cb5e8b5b-0513-4392-a079-0dab7664ec84\" (UID: \"cb5e8b5b-0513-4392-a079-0dab7664ec84\") " Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.824086 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb5e8b5b-0513-4392-a079-0dab7664ec84-config-data\") pod \"cb5e8b5b-0513-4392-a079-0dab7664ec84\" (UID: \"cb5e8b5b-0513-4392-a079-0dab7664ec84\") " Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.824186 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb5e8b5b-0513-4392-a079-0dab7664ec84-config-data-custom\") pod \"cb5e8b5b-0513-4392-a079-0dab7664ec84\" (UID: \"cb5e8b5b-0513-4392-a079-0dab7664ec84\") " Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.824229 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mrcf\" (UniqueName: \"kubernetes.io/projected/cb5e8b5b-0513-4392-a079-0dab7664ec84-kube-api-access-2mrcf\") pod \"cb5e8b5b-0513-4392-a079-0dab7664ec84\" (UID: \"cb5e8b5b-0513-4392-a079-0dab7664ec84\") " Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.824264 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb5e8b5b-0513-4392-a079-0dab7664ec84-etc-machine-id\") pod \"cb5e8b5b-0513-4392-a079-0dab7664ec84\" (UID: \"cb5e8b5b-0513-4392-a079-0dab7664ec84\") " Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.824300 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb5e8b5b-0513-4392-a079-0dab7664ec84-combined-ca-bundle\") pod \"cb5e8b5b-0513-4392-a079-0dab7664ec84\" (UID: \"cb5e8b5b-0513-4392-a079-0dab7664ec84\") " Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.824827 4559 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b649ece4-50a8-4449-ba98-2815000866fa-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.824841 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/b649ece4-50a8-4449-ba98-2815000866fa-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.824850 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4f62\" (UniqueName: \"kubernetes.io/projected/b649ece4-50a8-4449-ba98-2815000866fa-kube-api-access-t4f62\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.824861 4559 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b649ece4-50a8-4449-ba98-2815000866fa-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.824871 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b649ece4-50a8-4449-ba98-2815000866fa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.829411 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb5e8b5b-0513-4392-a079-0dab7664ec84-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "cb5e8b5b-0513-4392-a079-0dab7664ec84" (UID: "cb5e8b5b-0513-4392-a079-0dab7664ec84"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.831441 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cb5e8b5b-0513-4392-a079-0dab7664ec84-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "cb5e8b5b-0513-4392-a079-0dab7664ec84" (UID: "cb5e8b5b-0513-4392-a079-0dab7664ec84"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.832832 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb5e8b5b-0513-4392-a079-0dab7664ec84-scripts" (OuterVolumeSpecName: "scripts") pod "cb5e8b5b-0513-4392-a079-0dab7664ec84" (UID: "cb5e8b5b-0513-4392-a079-0dab7664ec84"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.834726 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb5e8b5b-0513-4392-a079-0dab7664ec84-kube-api-access-2mrcf" (OuterVolumeSpecName: "kube-api-access-2mrcf") pod "cb5e8b5b-0513-4392-a079-0dab7664ec84" (UID: "cb5e8b5b-0513-4392-a079-0dab7664ec84"). InnerVolumeSpecName "kube-api-access-2mrcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.869843 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb5e8b5b-0513-4392-a079-0dab7664ec84-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb5e8b5b-0513-4392-a079-0dab7664ec84" (UID: "cb5e8b5b-0513-4392-a079-0dab7664ec84"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.918868 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb5e8b5b-0513-4392-a079-0dab7664ec84-config-data" (OuterVolumeSpecName: "config-data") pod "cb5e8b5b-0513-4392-a079-0dab7664ec84" (UID: "cb5e8b5b-0513-4392-a079-0dab7664ec84"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.926577 4559 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb5e8b5b-0513-4392-a079-0dab7664ec84-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.926607 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mrcf\" (UniqueName: \"kubernetes.io/projected/cb5e8b5b-0513-4392-a079-0dab7664ec84-kube-api-access-2mrcf\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.926619 4559 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb5e8b5b-0513-4392-a079-0dab7664ec84-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.926629 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb5e8b5b-0513-4392-a079-0dab7664ec84-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.926657 4559 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb5e8b5b-0513-4392-a079-0dab7664ec84-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:12 crc kubenswrapper[4559]: I1123 07:00:12.926665 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb5e8b5b-0513-4392-a079-0dab7664ec84-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.197611 4559 generic.go:334] "Generic (PLEG): container finished" podID="cb5e8b5b-0513-4392-a079-0dab7664ec84" containerID="1d0ef4f29080e8dba7e9ab6b9fc27101b075c4bc0291cf4f4b8f734983e8543a" exitCode=0 Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.197709 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb5e8b5b-0513-4392-a079-0dab7664ec84","Type":"ContainerDied","Data":"1d0ef4f29080e8dba7e9ab6b9fc27101b075c4bc0291cf4f4b8f734983e8543a"} Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.197741 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb5e8b5b-0513-4392-a079-0dab7664ec84","Type":"ContainerDied","Data":"17e7bb197c8a3788374cdec264eb06fb2f9ec9e587d6787d61eb898d47efa22a"} Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.197760 4559 scope.go:117] "RemoveContainer" containerID="704862b0ff2fe684d984c817f5989629201f75ed385a420513da5c764edd5a15" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.197865 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.200429 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67f8477cdd-4x7mh" event={"ID":"b649ece4-50a8-4449-ba98-2815000866fa","Type":"ContainerDied","Data":"20e1977fa028bc9ca61faf1d37d02151b2d4a56fa46d18a3c1edf628df12f3a7"} Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.200713 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-67f8477cdd-4x7mh" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.218610 4559 scope.go:117] "RemoveContainer" containerID="1d0ef4f29080e8dba7e9ab6b9fc27101b075c4bc0291cf4f4b8f734983e8543a" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.239739 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-67f8477cdd-4x7mh"] Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.246134 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-67f8477cdd-4x7mh"] Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.290233 4559 scope.go:117] "RemoveContainer" containerID="704862b0ff2fe684d984c817f5989629201f75ed385a420513da5c764edd5a15" Nov 23 07:00:13 crc kubenswrapper[4559]: E1123 07:00:13.290760 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"704862b0ff2fe684d984c817f5989629201f75ed385a420513da5c764edd5a15\": container with ID starting with 704862b0ff2fe684d984c817f5989629201f75ed385a420513da5c764edd5a15 not found: ID does not exist" containerID="704862b0ff2fe684d984c817f5989629201f75ed385a420513da5c764edd5a15" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.290803 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"704862b0ff2fe684d984c817f5989629201f75ed385a420513da5c764edd5a15"} err="failed to get container status \"704862b0ff2fe684d984c817f5989629201f75ed385a420513da5c764edd5a15\": rpc error: code = NotFound desc = could not find container \"704862b0ff2fe684d984c817f5989629201f75ed385a420513da5c764edd5a15\": container with ID starting with 704862b0ff2fe684d984c817f5989629201f75ed385a420513da5c764edd5a15 not found: ID does not exist" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.290829 4559 scope.go:117] "RemoveContainer" containerID="1d0ef4f29080e8dba7e9ab6b9fc27101b075c4bc0291cf4f4b8f734983e8543a" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.291500 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 07:00:13 crc kubenswrapper[4559]: E1123 07:00:13.291532 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d0ef4f29080e8dba7e9ab6b9fc27101b075c4bc0291cf4f4b8f734983e8543a\": container with ID starting with 1d0ef4f29080e8dba7e9ab6b9fc27101b075c4bc0291cf4f4b8f734983e8543a not found: ID does not exist" containerID="1d0ef4f29080e8dba7e9ab6b9fc27101b075c4bc0291cf4f4b8f734983e8543a" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.291564 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d0ef4f29080e8dba7e9ab6b9fc27101b075c4bc0291cf4f4b8f734983e8543a"} err="failed to get container status \"1d0ef4f29080e8dba7e9ab6b9fc27101b075c4bc0291cf4f4b8f734983e8543a\": rpc error: code = NotFound desc = could not find container \"1d0ef4f29080e8dba7e9ab6b9fc27101b075c4bc0291cf4f4b8f734983e8543a\": container with ID starting with 1d0ef4f29080e8dba7e9ab6b9fc27101b075c4bc0291cf4f4b8f734983e8543a not found: ID does not exist" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.291589 4559 scope.go:117] "RemoveContainer" containerID="70a6af06589557c859bf0b9b1efe59d4f712ca7df817304e61ba10633adc6a9b" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.306849 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.319991 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 07:00:13 crc kubenswrapper[4559]: E1123 07:00:13.320356 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b649ece4-50a8-4449-ba98-2815000866fa" containerName="neutron-httpd" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.320374 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="b649ece4-50a8-4449-ba98-2815000866fa" containerName="neutron-httpd" Nov 23 07:00:13 crc kubenswrapper[4559]: E1123 07:00:13.320392 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69cec8df-ef13-4a2a-884b-78ba11c34b3e" containerName="barbican-api-log" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.320398 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="69cec8df-ef13-4a2a-884b-78ba11c34b3e" containerName="barbican-api-log" Nov 23 07:00:13 crc kubenswrapper[4559]: E1123 07:00:13.320406 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69cec8df-ef13-4a2a-884b-78ba11c34b3e" containerName="barbican-api" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.320412 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="69cec8df-ef13-4a2a-884b-78ba11c34b3e" containerName="barbican-api" Nov 23 07:00:13 crc kubenswrapper[4559]: E1123 07:00:13.320422 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b649ece4-50a8-4449-ba98-2815000866fa" containerName="neutron-api" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.320427 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="b649ece4-50a8-4449-ba98-2815000866fa" containerName="neutron-api" Nov 23 07:00:13 crc kubenswrapper[4559]: E1123 07:00:13.320443 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb5e8b5b-0513-4392-a079-0dab7664ec84" containerName="probe" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.320450 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb5e8b5b-0513-4392-a079-0dab7664ec84" containerName="probe" Nov 23 07:00:13 crc kubenswrapper[4559]: E1123 07:00:13.320474 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0" containerName="init" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.320480 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0" containerName="init" Nov 23 07:00:13 crc kubenswrapper[4559]: E1123 07:00:13.320488 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0" containerName="dnsmasq-dns" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.320494 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0" containerName="dnsmasq-dns" Nov 23 07:00:13 crc kubenswrapper[4559]: E1123 07:00:13.320501 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb5e8b5b-0513-4392-a079-0dab7664ec84" containerName="cinder-scheduler" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.320506 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb5e8b5b-0513-4392-a079-0dab7664ec84" containerName="cinder-scheduler" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.320745 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="b649ece4-50a8-4449-ba98-2815000866fa" containerName="neutron-api" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.320770 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fb0ec1c-da7b-48a9-a83a-3bfc410b0ba0" containerName="dnsmasq-dns" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.320776 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="69cec8df-ef13-4a2a-884b-78ba11c34b3e" containerName="barbican-api" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.320786 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb5e8b5b-0513-4392-a079-0dab7664ec84" containerName="probe" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.320794 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="b649ece4-50a8-4449-ba98-2815000866fa" containerName="neutron-httpd" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.320808 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb5e8b5b-0513-4392-a079-0dab7664ec84" containerName="cinder-scheduler" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.320818 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="69cec8df-ef13-4a2a-884b-78ba11c34b3e" containerName="barbican-api-log" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.321102 4559 scope.go:117] "RemoveContainer" containerID="bd5950999e0816dc3a6c6c59ae79a87d4896d498e6aec37ee3fb20cea46e18c3" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.321711 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.325325 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.328214 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.442396 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd7f1662-fa90-4677-9c1a-e5882f079497-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cd7f1662-fa90-4677-9c1a-e5882f079497\") " pod="openstack/cinder-scheduler-0" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.442544 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd7f1662-fa90-4677-9c1a-e5882f079497-config-data\") pod \"cinder-scheduler-0\" (UID: \"cd7f1662-fa90-4677-9c1a-e5882f079497\") " pod="openstack/cinder-scheduler-0" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.442665 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q74xr\" (UniqueName: \"kubernetes.io/projected/cd7f1662-fa90-4677-9c1a-e5882f079497-kube-api-access-q74xr\") pod \"cinder-scheduler-0\" (UID: \"cd7f1662-fa90-4677-9c1a-e5882f079497\") " pod="openstack/cinder-scheduler-0" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.442709 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd7f1662-fa90-4677-9c1a-e5882f079497-scripts\") pod \"cinder-scheduler-0\" (UID: \"cd7f1662-fa90-4677-9c1a-e5882f079497\") " pod="openstack/cinder-scheduler-0" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.442724 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cd7f1662-fa90-4677-9c1a-e5882f079497-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cd7f1662-fa90-4677-9c1a-e5882f079497\") " pod="openstack/cinder-scheduler-0" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.442747 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd7f1662-fa90-4677-9c1a-e5882f079497-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cd7f1662-fa90-4677-9c1a-e5882f079497\") " pod="openstack/cinder-scheduler-0" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.544352 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q74xr\" (UniqueName: \"kubernetes.io/projected/cd7f1662-fa90-4677-9c1a-e5882f079497-kube-api-access-q74xr\") pod \"cinder-scheduler-0\" (UID: \"cd7f1662-fa90-4677-9c1a-e5882f079497\") " pod="openstack/cinder-scheduler-0" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.544399 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd7f1662-fa90-4677-9c1a-e5882f079497-scripts\") pod \"cinder-scheduler-0\" (UID: \"cd7f1662-fa90-4677-9c1a-e5882f079497\") " pod="openstack/cinder-scheduler-0" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.544419 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cd7f1662-fa90-4677-9c1a-e5882f079497-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cd7f1662-fa90-4677-9c1a-e5882f079497\") " pod="openstack/cinder-scheduler-0" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.544438 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd7f1662-fa90-4677-9c1a-e5882f079497-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cd7f1662-fa90-4677-9c1a-e5882f079497\") " pod="openstack/cinder-scheduler-0" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.544458 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd7f1662-fa90-4677-9c1a-e5882f079497-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cd7f1662-fa90-4677-9c1a-e5882f079497\") " pod="openstack/cinder-scheduler-0" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.544557 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd7f1662-fa90-4677-9c1a-e5882f079497-config-data\") pod \"cinder-scheduler-0\" (UID: \"cd7f1662-fa90-4677-9c1a-e5882f079497\") " pod="openstack/cinder-scheduler-0" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.544548 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cd7f1662-fa90-4677-9c1a-e5882f079497-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cd7f1662-fa90-4677-9c1a-e5882f079497\") " pod="openstack/cinder-scheduler-0" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.551046 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd7f1662-fa90-4677-9c1a-e5882f079497-scripts\") pod \"cinder-scheduler-0\" (UID: \"cd7f1662-fa90-4677-9c1a-e5882f079497\") " pod="openstack/cinder-scheduler-0" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.551590 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd7f1662-fa90-4677-9c1a-e5882f079497-config-data\") pod \"cinder-scheduler-0\" (UID: \"cd7f1662-fa90-4677-9c1a-e5882f079497\") " pod="openstack/cinder-scheduler-0" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.553570 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd7f1662-fa90-4677-9c1a-e5882f079497-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cd7f1662-fa90-4677-9c1a-e5882f079497\") " pod="openstack/cinder-scheduler-0" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.556428 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd7f1662-fa90-4677-9c1a-e5882f079497-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cd7f1662-fa90-4677-9c1a-e5882f079497\") " pod="openstack/cinder-scheduler-0" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.561852 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q74xr\" (UniqueName: \"kubernetes.io/projected/cd7f1662-fa90-4677-9c1a-e5882f079497-kube-api-access-q74xr\") pod \"cinder-scheduler-0\" (UID: \"cd7f1662-fa90-4677-9c1a-e5882f079497\") " pod="openstack/cinder-scheduler-0" Nov 23 07:00:13 crc kubenswrapper[4559]: I1123 07:00:13.646409 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 23 07:00:14 crc kubenswrapper[4559]: I1123 07:00:14.058092 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 07:00:14 crc kubenswrapper[4559]: W1123 07:00:14.058572 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd7f1662_fa90_4677_9c1a_e5882f079497.slice/crio-32f026603da3ad0e03555a0116ad3b48143ec6ce78fe1c6f163898455a6e8d8b WatchSource:0}: Error finding container 32f026603da3ad0e03555a0116ad3b48143ec6ce78fe1c6f163898455a6e8d8b: Status 404 returned error can't find the container with id 32f026603da3ad0e03555a0116ad3b48143ec6ce78fe1c6f163898455a6e8d8b Nov 23 07:00:14 crc kubenswrapper[4559]: I1123 07:00:14.222780 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cd7f1662-fa90-4677-9c1a-e5882f079497","Type":"ContainerStarted","Data":"32f026603da3ad0e03555a0116ad3b48143ec6ce78fe1c6f163898455a6e8d8b"} Nov 23 07:00:14 crc kubenswrapper[4559]: I1123 07:00:14.286181 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b649ece4-50a8-4449-ba98-2815000866fa" path="/var/lib/kubelet/pods/b649ece4-50a8-4449-ba98-2815000866fa/volumes" Nov 23 07:00:14 crc kubenswrapper[4559]: I1123 07:00:14.286809 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb5e8b5b-0513-4392-a079-0dab7664ec84" path="/var/lib/kubelet/pods/cb5e8b5b-0513-4392-a079-0dab7664ec84/volumes" Nov 23 07:00:15 crc kubenswrapper[4559]: I1123 07:00:15.066506 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 23 07:00:15 crc kubenswrapper[4559]: I1123 07:00:15.263336 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cd7f1662-fa90-4677-9c1a-e5882f079497","Type":"ContainerStarted","Data":"85afc659e3bcc68bac63c0055b242e1e2c3e38412bfb453d57f42b36aae6c123"} Nov 23 07:00:15 crc kubenswrapper[4559]: I1123 07:00:15.263378 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cd7f1662-fa90-4677-9c1a-e5882f079497","Type":"ContainerStarted","Data":"9def1d2be14e5ef867089c6f08d678b066fab69e7da90f3f30a5317bcb0c2072"} Nov 23 07:00:18 crc kubenswrapper[4559]: I1123 07:00:18.646491 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 23 07:00:21 crc kubenswrapper[4559]: I1123 07:00:21.443962 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-74ff86f86c-gcbcl" Nov 23 07:00:21 crc kubenswrapper[4559]: I1123 07:00:21.461499 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=8.461478057 podStartE2EDuration="8.461478057s" podCreationTimestamp="2025-11-23 07:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:00:15.292862339 +0000 UTC m=+917.314847953" watchObservedRunningTime="2025-11-23 07:00:21.461478057 +0000 UTC m=+923.483463672" Nov 23 07:00:23 crc kubenswrapper[4559]: I1123 07:00:23.828067 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.041059 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.050246 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.050362 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.052177 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.053621 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-wtrhb" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.053826 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.143905 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a0d11d6f-21b2-4db8-b48f-a14c90f52860-openstack-config\") pod \"openstackclient\" (UID: \"a0d11d6f-21b2-4db8-b48f-a14c90f52860\") " pod="openstack/openstackclient" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.143962 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a0d11d6f-21b2-4db8-b48f-a14c90f52860-openstack-config-secret\") pod \"openstackclient\" (UID: \"a0d11d6f-21b2-4db8-b48f-a14c90f52860\") " pod="openstack/openstackclient" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.143995 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0d11d6f-21b2-4db8-b48f-a14c90f52860-combined-ca-bundle\") pod \"openstackclient\" (UID: \"a0d11d6f-21b2-4db8-b48f-a14c90f52860\") " pod="openstack/openstackclient" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.144054 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sfrc\" (UniqueName: \"kubernetes.io/projected/a0d11d6f-21b2-4db8-b48f-a14c90f52860-kube-api-access-6sfrc\") pod \"openstackclient\" (UID: \"a0d11d6f-21b2-4db8-b48f-a14c90f52860\") " pod="openstack/openstackclient" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.245409 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a0d11d6f-21b2-4db8-b48f-a14c90f52860-openstack-config\") pod \"openstackclient\" (UID: \"a0d11d6f-21b2-4db8-b48f-a14c90f52860\") " pod="openstack/openstackclient" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.245450 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a0d11d6f-21b2-4db8-b48f-a14c90f52860-openstack-config-secret\") pod \"openstackclient\" (UID: \"a0d11d6f-21b2-4db8-b48f-a14c90f52860\") " pod="openstack/openstackclient" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.245475 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0d11d6f-21b2-4db8-b48f-a14c90f52860-combined-ca-bundle\") pod \"openstackclient\" (UID: \"a0d11d6f-21b2-4db8-b48f-a14c90f52860\") " pod="openstack/openstackclient" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.245532 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sfrc\" (UniqueName: \"kubernetes.io/projected/a0d11d6f-21b2-4db8-b48f-a14c90f52860-kube-api-access-6sfrc\") pod \"openstackclient\" (UID: \"a0d11d6f-21b2-4db8-b48f-a14c90f52860\") " pod="openstack/openstackclient" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.246531 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a0d11d6f-21b2-4db8-b48f-a14c90f52860-openstack-config\") pod \"openstackclient\" (UID: \"a0d11d6f-21b2-4db8-b48f-a14c90f52860\") " pod="openstack/openstackclient" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.252075 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a0d11d6f-21b2-4db8-b48f-a14c90f52860-openstack-config-secret\") pod \"openstackclient\" (UID: \"a0d11d6f-21b2-4db8-b48f-a14c90f52860\") " pod="openstack/openstackclient" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.252321 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0d11d6f-21b2-4db8-b48f-a14c90f52860-combined-ca-bundle\") pod \"openstackclient\" (UID: \"a0d11d6f-21b2-4db8-b48f-a14c90f52860\") " pod="openstack/openstackclient" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.259828 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sfrc\" (UniqueName: \"kubernetes.io/projected/a0d11d6f-21b2-4db8-b48f-a14c90f52860-kube-api-access-6sfrc\") pod \"openstackclient\" (UID: \"a0d11d6f-21b2-4db8-b48f-a14c90f52860\") " pod="openstack/openstackclient" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.374968 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.422168 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.428764 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.452225 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.453617 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.458557 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 23 07:00:24 crc kubenswrapper[4559]: E1123 07:00:24.494530 4559 log.go:32] "RunPodSandbox from runtime service failed" err=< Nov 23 07:00:24 crc kubenswrapper[4559]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_a0d11d6f-21b2-4db8-b48f-a14c90f52860_0(dd88200f69adb44d08ae52c7c2ee95053662f9f95c352e1526617cf1a8b5e600): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"dd88200f69adb44d08ae52c7c2ee95053662f9f95c352e1526617cf1a8b5e600" Netns:"/var/run/netns/f88cb266-023d-4900-8fe8-52c71c113bc9" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=dd88200f69adb44d08ae52c7c2ee95053662f9f95c352e1526617cf1a8b5e600;K8S_POD_UID=a0d11d6f-21b2-4db8-b48f-a14c90f52860" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/a0d11d6f-21b2-4db8-b48f-a14c90f52860]: expected pod UID "a0d11d6f-21b2-4db8-b48f-a14c90f52860" but got "0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa" from Kube API Nov 23 07:00:24 crc kubenswrapper[4559]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Nov 23 07:00:24 crc kubenswrapper[4559]: > Nov 23 07:00:24 crc kubenswrapper[4559]: E1123 07:00:24.494599 4559 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Nov 23 07:00:24 crc kubenswrapper[4559]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_a0d11d6f-21b2-4db8-b48f-a14c90f52860_0(dd88200f69adb44d08ae52c7c2ee95053662f9f95c352e1526617cf1a8b5e600): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"dd88200f69adb44d08ae52c7c2ee95053662f9f95c352e1526617cf1a8b5e600" Netns:"/var/run/netns/f88cb266-023d-4900-8fe8-52c71c113bc9" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=dd88200f69adb44d08ae52c7c2ee95053662f9f95c352e1526617cf1a8b5e600;K8S_POD_UID=a0d11d6f-21b2-4db8-b48f-a14c90f52860" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/a0d11d6f-21b2-4db8-b48f-a14c90f52860]: expected pod UID "a0d11d6f-21b2-4db8-b48f-a14c90f52860" but got "0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa" from Kube API Nov 23 07:00:24 crc kubenswrapper[4559]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Nov 23 07:00:24 crc kubenswrapper[4559]: > pod="openstack/openstackclient" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.554526 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa\") " pod="openstack/openstackclient" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.554592 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa-openstack-config\") pod \"openstackclient\" (UID: \"0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa\") " pod="openstack/openstackclient" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.554798 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jlbl\" (UniqueName: \"kubernetes.io/projected/0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa-kube-api-access-4jlbl\") pod \"openstackclient\" (UID: \"0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa\") " pod="openstack/openstackclient" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.555024 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa-openstack-config-secret\") pod \"openstackclient\" (UID: \"0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa\") " pod="openstack/openstackclient" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.656079 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jlbl\" (UniqueName: \"kubernetes.io/projected/0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa-kube-api-access-4jlbl\") pod \"openstackclient\" (UID: \"0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa\") " pod="openstack/openstackclient" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.656182 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa-openstack-config-secret\") pod \"openstackclient\" (UID: \"0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa\") " pod="openstack/openstackclient" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.656233 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa\") " pod="openstack/openstackclient" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.656262 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa-openstack-config\") pod \"openstackclient\" (UID: \"0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa\") " pod="openstack/openstackclient" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.657138 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa-openstack-config\") pod \"openstackclient\" (UID: \"0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa\") " pod="openstack/openstackclient" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.661964 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa-openstack-config-secret\") pod \"openstackclient\" (UID: \"0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa\") " pod="openstack/openstackclient" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.662256 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa\") " pod="openstack/openstackclient" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.669386 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jlbl\" (UniqueName: \"kubernetes.io/projected/0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa-kube-api-access-4jlbl\") pod \"openstackclient\" (UID: \"0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa\") " pod="openstack/openstackclient" Nov 23 07:00:24 crc kubenswrapper[4559]: I1123 07:00:24.831740 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 23 07:00:25 crc kubenswrapper[4559]: I1123 07:00:25.309485 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 23 07:00:25 crc kubenswrapper[4559]: I1123 07:00:25.371842 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa","Type":"ContainerStarted","Data":"64e5ff7f758c4d209883ae962fb593ed220dae8f1cbd7510cea92ccf76755675"} Nov 23 07:00:25 crc kubenswrapper[4559]: I1123 07:00:25.371908 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 23 07:00:25 crc kubenswrapper[4559]: I1123 07:00:25.374764 4559 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="a0d11d6f-21b2-4db8-b48f-a14c90f52860" podUID="0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa" Nov 23 07:00:25 crc kubenswrapper[4559]: I1123 07:00:25.382151 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 23 07:00:25 crc kubenswrapper[4559]: I1123 07:00:25.480050 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a0d11d6f-21b2-4db8-b48f-a14c90f52860-openstack-config-secret\") pod \"a0d11d6f-21b2-4db8-b48f-a14c90f52860\" (UID: \"a0d11d6f-21b2-4db8-b48f-a14c90f52860\") " Nov 23 07:00:25 crc kubenswrapper[4559]: I1123 07:00:25.480150 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sfrc\" (UniqueName: \"kubernetes.io/projected/a0d11d6f-21b2-4db8-b48f-a14c90f52860-kube-api-access-6sfrc\") pod \"a0d11d6f-21b2-4db8-b48f-a14c90f52860\" (UID: \"a0d11d6f-21b2-4db8-b48f-a14c90f52860\") " Nov 23 07:00:25 crc kubenswrapper[4559]: I1123 07:00:25.480257 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0d11d6f-21b2-4db8-b48f-a14c90f52860-combined-ca-bundle\") pod \"a0d11d6f-21b2-4db8-b48f-a14c90f52860\" (UID: \"a0d11d6f-21b2-4db8-b48f-a14c90f52860\") " Nov 23 07:00:25 crc kubenswrapper[4559]: I1123 07:00:25.480299 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a0d11d6f-21b2-4db8-b48f-a14c90f52860-openstack-config\") pod \"a0d11d6f-21b2-4db8-b48f-a14c90f52860\" (UID: \"a0d11d6f-21b2-4db8-b48f-a14c90f52860\") " Nov 23 07:00:25 crc kubenswrapper[4559]: I1123 07:00:25.481261 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0d11d6f-21b2-4db8-b48f-a14c90f52860-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "a0d11d6f-21b2-4db8-b48f-a14c90f52860" (UID: "a0d11d6f-21b2-4db8-b48f-a14c90f52860"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:00:25 crc kubenswrapper[4559]: I1123 07:00:25.496361 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0d11d6f-21b2-4db8-b48f-a14c90f52860-kube-api-access-6sfrc" (OuterVolumeSpecName: "kube-api-access-6sfrc") pod "a0d11d6f-21b2-4db8-b48f-a14c90f52860" (UID: "a0d11d6f-21b2-4db8-b48f-a14c90f52860"). InnerVolumeSpecName "kube-api-access-6sfrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:00:25 crc kubenswrapper[4559]: I1123 07:00:25.498726 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0d11d6f-21b2-4db8-b48f-a14c90f52860-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "a0d11d6f-21b2-4db8-b48f-a14c90f52860" (UID: "a0d11d6f-21b2-4db8-b48f-a14c90f52860"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:25 crc kubenswrapper[4559]: I1123 07:00:25.498823 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0d11d6f-21b2-4db8-b48f-a14c90f52860-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a0d11d6f-21b2-4db8-b48f-a14c90f52860" (UID: "a0d11d6f-21b2-4db8-b48f-a14c90f52860"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:25 crc kubenswrapper[4559]: I1123 07:00:25.582811 4559 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a0d11d6f-21b2-4db8-b48f-a14c90f52860-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:25 crc kubenswrapper[4559]: I1123 07:00:25.582842 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sfrc\" (UniqueName: \"kubernetes.io/projected/a0d11d6f-21b2-4db8-b48f-a14c90f52860-kube-api-access-6sfrc\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:25 crc kubenswrapper[4559]: I1123 07:00:25.582853 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0d11d6f-21b2-4db8-b48f-a14c90f52860-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:25 crc kubenswrapper[4559]: I1123 07:00:25.582862 4559 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a0d11d6f-21b2-4db8-b48f-a14c90f52860-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:26 crc kubenswrapper[4559]: I1123 07:00:26.167006 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:00:26 crc kubenswrapper[4559]: I1123 07:00:26.167084 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:00:26 crc kubenswrapper[4559]: I1123 07:00:26.167143 4559 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" Nov 23 07:00:26 crc kubenswrapper[4559]: I1123 07:00:26.168680 4559 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"667a247ddea953ad2f29e458b8a21398c7c8c20545d619492b9346eff264de8f"} pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 07:00:26 crc kubenswrapper[4559]: I1123 07:00:26.168767 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" containerID="cri-o://667a247ddea953ad2f29e458b8a21398c7c8c20545d619492b9346eff264de8f" gracePeriod=600 Nov 23 07:00:26 crc kubenswrapper[4559]: I1123 07:00:26.284756 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0d11d6f-21b2-4db8-b48f-a14c90f52860" path="/var/lib/kubelet/pods/a0d11d6f-21b2-4db8-b48f-a14c90f52860/volumes" Nov 23 07:00:26 crc kubenswrapper[4559]: I1123 07:00:26.391130 4559 generic.go:334] "Generic (PLEG): container finished" podID="4731beee-0cac-4189-8a70-743b0b709095" containerID="667a247ddea953ad2f29e458b8a21398c7c8c20545d619492b9346eff264de8f" exitCode=0 Nov 23 07:00:26 crc kubenswrapper[4559]: I1123 07:00:26.391200 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 23 07:00:26 crc kubenswrapper[4559]: I1123 07:00:26.391572 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerDied","Data":"667a247ddea953ad2f29e458b8a21398c7c8c20545d619492b9346eff264de8f"} Nov 23 07:00:26 crc kubenswrapper[4559]: I1123 07:00:26.391606 4559 scope.go:117] "RemoveContainer" containerID="c1a47810490d41ffdaf229d9abf61dc047a9840ab17302f7f8ed9bcafa8fe6db" Nov 23 07:00:26 crc kubenswrapper[4559]: I1123 07:00:26.426280 4559 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="a0d11d6f-21b2-4db8-b48f-a14c90f52860" podUID="0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa" Nov 23 07:00:27 crc kubenswrapper[4559]: I1123 07:00:27.409472 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerStarted","Data":"b7be9dbebcbc9b258e6fc583ce1dee8a3804425c02033654afb5190f533977ae"} Nov 23 07:00:28 crc kubenswrapper[4559]: I1123 07:00:28.138996 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:00:28 crc kubenswrapper[4559]: I1123 07:00:28.139513 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b0ee294-003c-49b1-878e-50097454795f" containerName="ceilometer-central-agent" containerID="cri-o://c21b860c45243f8824cd1b77c0a54694e67b445eac3577a219d0e5d4f7e8ce15" gracePeriod=30 Nov 23 07:00:28 crc kubenswrapper[4559]: I1123 07:00:28.139573 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b0ee294-003c-49b1-878e-50097454795f" containerName="proxy-httpd" containerID="cri-o://81ccb4b66519f8f7d2facac1d0226906cf29314b35e5f091f2c049fa273ef2e2" gracePeriod=30 Nov 23 07:00:28 crc kubenswrapper[4559]: I1123 07:00:28.139573 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b0ee294-003c-49b1-878e-50097454795f" containerName="sg-core" containerID="cri-o://ae1b972ec5cc95c188cb0fac2c7e0ec994b95f2fa2655a84967d824390d1ceb4" gracePeriod=30 Nov 23 07:00:28 crc kubenswrapper[4559]: I1123 07:00:28.139704 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b0ee294-003c-49b1-878e-50097454795f" containerName="ceilometer-notification-agent" containerID="cri-o://46666f0629352c2d81b02e541670d6640a217e0d5b21a838ed83703abd0d91ef" gracePeriod=30 Nov 23 07:00:28 crc kubenswrapper[4559]: I1123 07:00:28.146968 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 23 07:00:28 crc kubenswrapper[4559]: I1123 07:00:28.425272 4559 generic.go:334] "Generic (PLEG): container finished" podID="5b0ee294-003c-49b1-878e-50097454795f" containerID="81ccb4b66519f8f7d2facac1d0226906cf29314b35e5f091f2c049fa273ef2e2" exitCode=0 Nov 23 07:00:28 crc kubenswrapper[4559]: I1123 07:00:28.425327 4559 generic.go:334] "Generic (PLEG): container finished" podID="5b0ee294-003c-49b1-878e-50097454795f" containerID="ae1b972ec5cc95c188cb0fac2c7e0ec994b95f2fa2655a84967d824390d1ceb4" exitCode=2 Nov 23 07:00:28 crc kubenswrapper[4559]: I1123 07:00:28.425872 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b0ee294-003c-49b1-878e-50097454795f","Type":"ContainerDied","Data":"81ccb4b66519f8f7d2facac1d0226906cf29314b35e5f091f2c049fa273ef2e2"} Nov 23 07:00:28 crc kubenswrapper[4559]: I1123 07:00:28.425927 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b0ee294-003c-49b1-878e-50097454795f","Type":"ContainerDied","Data":"ae1b972ec5cc95c188cb0fac2c7e0ec994b95f2fa2655a84967d824390d1ceb4"} Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.259072 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-6b4fdd88f9-9hnng"] Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.261100 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.264244 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.264413 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.264577 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.280495 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6b4fdd88f9-9hnng"] Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.361214 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/db083799-cb94-42d3-8a54-8446f7a76502-public-tls-certs\") pod \"swift-proxy-6b4fdd88f9-9hnng\" (UID: \"db083799-cb94-42d3-8a54-8446f7a76502\") " pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.361260 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwwxd\" (UniqueName: \"kubernetes.io/projected/db083799-cb94-42d3-8a54-8446f7a76502-kube-api-access-wwwxd\") pod \"swift-proxy-6b4fdd88f9-9hnng\" (UID: \"db083799-cb94-42d3-8a54-8446f7a76502\") " pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.361341 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db083799-cb94-42d3-8a54-8446f7a76502-run-httpd\") pod \"swift-proxy-6b4fdd88f9-9hnng\" (UID: \"db083799-cb94-42d3-8a54-8446f7a76502\") " pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.361411 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db083799-cb94-42d3-8a54-8446f7a76502-combined-ca-bundle\") pod \"swift-proxy-6b4fdd88f9-9hnng\" (UID: \"db083799-cb94-42d3-8a54-8446f7a76502\") " pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.361466 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/db083799-cb94-42d3-8a54-8446f7a76502-internal-tls-certs\") pod \"swift-proxy-6b4fdd88f9-9hnng\" (UID: \"db083799-cb94-42d3-8a54-8446f7a76502\") " pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.361495 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/db083799-cb94-42d3-8a54-8446f7a76502-etc-swift\") pod \"swift-proxy-6b4fdd88f9-9hnng\" (UID: \"db083799-cb94-42d3-8a54-8446f7a76502\") " pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.361686 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db083799-cb94-42d3-8a54-8446f7a76502-config-data\") pod \"swift-proxy-6b4fdd88f9-9hnng\" (UID: \"db083799-cb94-42d3-8a54-8446f7a76502\") " pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.361852 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db083799-cb94-42d3-8a54-8446f7a76502-log-httpd\") pod \"swift-proxy-6b4fdd88f9-9hnng\" (UID: \"db083799-cb94-42d3-8a54-8446f7a76502\") " pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.434332 4559 generic.go:334] "Generic (PLEG): container finished" podID="5b0ee294-003c-49b1-878e-50097454795f" containerID="c21b860c45243f8824cd1b77c0a54694e67b445eac3577a219d0e5d4f7e8ce15" exitCode=0 Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.434379 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b0ee294-003c-49b1-878e-50097454795f","Type":"ContainerDied","Data":"c21b860c45243f8824cd1b77c0a54694e67b445eac3577a219d0e5d4f7e8ce15"} Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.463973 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/db083799-cb94-42d3-8a54-8446f7a76502-etc-swift\") pod \"swift-proxy-6b4fdd88f9-9hnng\" (UID: \"db083799-cb94-42d3-8a54-8446f7a76502\") " pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.464054 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db083799-cb94-42d3-8a54-8446f7a76502-config-data\") pod \"swift-proxy-6b4fdd88f9-9hnng\" (UID: \"db083799-cb94-42d3-8a54-8446f7a76502\") " pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.464118 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db083799-cb94-42d3-8a54-8446f7a76502-log-httpd\") pod \"swift-proxy-6b4fdd88f9-9hnng\" (UID: \"db083799-cb94-42d3-8a54-8446f7a76502\") " pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.464150 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/db083799-cb94-42d3-8a54-8446f7a76502-public-tls-certs\") pod \"swift-proxy-6b4fdd88f9-9hnng\" (UID: \"db083799-cb94-42d3-8a54-8446f7a76502\") " pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.464182 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwwxd\" (UniqueName: \"kubernetes.io/projected/db083799-cb94-42d3-8a54-8446f7a76502-kube-api-access-wwwxd\") pod \"swift-proxy-6b4fdd88f9-9hnng\" (UID: \"db083799-cb94-42d3-8a54-8446f7a76502\") " pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.464225 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db083799-cb94-42d3-8a54-8446f7a76502-run-httpd\") pod \"swift-proxy-6b4fdd88f9-9hnng\" (UID: \"db083799-cb94-42d3-8a54-8446f7a76502\") " pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.464259 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db083799-cb94-42d3-8a54-8446f7a76502-combined-ca-bundle\") pod \"swift-proxy-6b4fdd88f9-9hnng\" (UID: \"db083799-cb94-42d3-8a54-8446f7a76502\") " pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.464281 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/db083799-cb94-42d3-8a54-8446f7a76502-internal-tls-certs\") pod \"swift-proxy-6b4fdd88f9-9hnng\" (UID: \"db083799-cb94-42d3-8a54-8446f7a76502\") " pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.465501 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db083799-cb94-42d3-8a54-8446f7a76502-run-httpd\") pod \"swift-proxy-6b4fdd88f9-9hnng\" (UID: \"db083799-cb94-42d3-8a54-8446f7a76502\") " pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.468680 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db083799-cb94-42d3-8a54-8446f7a76502-log-httpd\") pod \"swift-proxy-6b4fdd88f9-9hnng\" (UID: \"db083799-cb94-42d3-8a54-8446f7a76502\") " pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.473552 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db083799-cb94-42d3-8a54-8446f7a76502-config-data\") pod \"swift-proxy-6b4fdd88f9-9hnng\" (UID: \"db083799-cb94-42d3-8a54-8446f7a76502\") " pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.473625 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db083799-cb94-42d3-8a54-8446f7a76502-combined-ca-bundle\") pod \"swift-proxy-6b4fdd88f9-9hnng\" (UID: \"db083799-cb94-42d3-8a54-8446f7a76502\") " pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.473718 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/db083799-cb94-42d3-8a54-8446f7a76502-internal-tls-certs\") pod \"swift-proxy-6b4fdd88f9-9hnng\" (UID: \"db083799-cb94-42d3-8a54-8446f7a76502\") " pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.473849 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/db083799-cb94-42d3-8a54-8446f7a76502-public-tls-certs\") pod \"swift-proxy-6b4fdd88f9-9hnng\" (UID: \"db083799-cb94-42d3-8a54-8446f7a76502\") " pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.474015 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/db083799-cb94-42d3-8a54-8446f7a76502-etc-swift\") pod \"swift-proxy-6b4fdd88f9-9hnng\" (UID: \"db083799-cb94-42d3-8a54-8446f7a76502\") " pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.488578 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwwxd\" (UniqueName: \"kubernetes.io/projected/db083799-cb94-42d3-8a54-8446f7a76502-kube-api-access-wwwxd\") pod \"swift-proxy-6b4fdd88f9-9hnng\" (UID: \"db083799-cb94-42d3-8a54-8446f7a76502\") " pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:29 crc kubenswrapper[4559]: I1123 07:00:29.591461 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:30 crc kubenswrapper[4559]: I1123 07:00:30.091614 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6b4fdd88f9-9hnng"] Nov 23 07:00:30 crc kubenswrapper[4559]: W1123 07:00:30.104798 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb083799_cb94_42d3_8a54_8446f7a76502.slice/crio-9392723799d25c82d7169d3de356cab693de06e5af07c0c77414daad1a5263fd WatchSource:0}: Error finding container 9392723799d25c82d7169d3de356cab693de06e5af07c0c77414daad1a5263fd: Status 404 returned error can't find the container with id 9392723799d25c82d7169d3de356cab693de06e5af07c0c77414daad1a5263fd Nov 23 07:00:30 crc kubenswrapper[4559]: I1123 07:00:30.444265 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6b4fdd88f9-9hnng" event={"ID":"db083799-cb94-42d3-8a54-8446f7a76502","Type":"ContainerStarted","Data":"1760ced01c880dea99c576308c1f52d402c8b758ada3a2633a762aa2cc162f01"} Nov 23 07:00:30 crc kubenswrapper[4559]: I1123 07:00:30.444540 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6b4fdd88f9-9hnng" event={"ID":"db083799-cb94-42d3-8a54-8446f7a76502","Type":"ContainerStarted","Data":"87f54bc799292ed56315057d192944a80066041f6b042d1becf48d693f1531f8"} Nov 23 07:00:30 crc kubenswrapper[4559]: I1123 07:00:30.444558 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:30 crc kubenswrapper[4559]: I1123 07:00:30.444570 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6b4fdd88f9-9hnng" event={"ID":"db083799-cb94-42d3-8a54-8446f7a76502","Type":"ContainerStarted","Data":"9392723799d25c82d7169d3de356cab693de06e5af07c0c77414daad1a5263fd"} Nov 23 07:00:30 crc kubenswrapper[4559]: I1123 07:00:30.469049 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-6b4fdd88f9-9hnng" podStartSLOduration=1.469030083 podStartE2EDuration="1.469030083s" podCreationTimestamp="2025-11-23 07:00:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:00:30.461984503 +0000 UTC m=+932.483970118" watchObservedRunningTime="2025-11-23 07:00:30.469030083 +0000 UTC m=+932.491015698" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.168027 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-t898j"] Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.169567 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-t898j" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.177118 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-t898j"] Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.199325 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stqxc\" (UniqueName: \"kubernetes.io/projected/ec53c68f-1637-47e8-8941-0ada991639ec-kube-api-access-stqxc\") pod \"nova-api-db-create-t898j\" (UID: \"ec53c68f-1637-47e8-8941-0ada991639ec\") " pod="openstack/nova-api-db-create-t898j" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.199462 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec53c68f-1637-47e8-8941-0ada991639ec-operator-scripts\") pod \"nova-api-db-create-t898j\" (UID: \"ec53c68f-1637-47e8-8941-0ada991639ec\") " pod="openstack/nova-api-db-create-t898j" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.262573 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-tp852"] Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.263818 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-tp852" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.275278 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-tp852"] Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.301856 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrjwv\" (UniqueName: \"kubernetes.io/projected/b1d3e297-8696-43ae-84f8-da003e5a03e5-kube-api-access-mrjwv\") pod \"nova-cell0-db-create-tp852\" (UID: \"b1d3e297-8696-43ae-84f8-da003e5a03e5\") " pod="openstack/nova-cell0-db-create-tp852" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.301970 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stqxc\" (UniqueName: \"kubernetes.io/projected/ec53c68f-1637-47e8-8941-0ada991639ec-kube-api-access-stqxc\") pod \"nova-api-db-create-t898j\" (UID: \"ec53c68f-1637-47e8-8941-0ada991639ec\") " pod="openstack/nova-api-db-create-t898j" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.302094 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec53c68f-1637-47e8-8941-0ada991639ec-operator-scripts\") pod \"nova-api-db-create-t898j\" (UID: \"ec53c68f-1637-47e8-8941-0ada991639ec\") " pod="openstack/nova-api-db-create-t898j" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.302120 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1d3e297-8696-43ae-84f8-da003e5a03e5-operator-scripts\") pod \"nova-cell0-db-create-tp852\" (UID: \"b1d3e297-8696-43ae-84f8-da003e5a03e5\") " pod="openstack/nova-cell0-db-create-tp852" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.303732 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec53c68f-1637-47e8-8941-0ada991639ec-operator-scripts\") pod \"nova-api-db-create-t898j\" (UID: \"ec53c68f-1637-47e8-8941-0ada991639ec\") " pod="openstack/nova-api-db-create-t898j" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.317735 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stqxc\" (UniqueName: \"kubernetes.io/projected/ec53c68f-1637-47e8-8941-0ada991639ec-kube-api-access-stqxc\") pod \"nova-api-db-create-t898j\" (UID: \"ec53c68f-1637-47e8-8941-0ada991639ec\") " pod="openstack/nova-api-db-create-t898j" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.370864 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-kktcm"] Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.372528 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kktcm" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.388451 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-d2d8-account-create-tfkjf"] Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.389809 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-d2d8-account-create-tfkjf" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.391067 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.398474 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-kktcm"] Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.403857 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztqns\" (UniqueName: \"kubernetes.io/projected/79afaafd-5f51-4862-9c09-e7f817d54f73-kube-api-access-ztqns\") pod \"nova-cell1-db-create-kktcm\" (UID: \"79afaafd-5f51-4862-9c09-e7f817d54f73\") " pod="openstack/nova-cell1-db-create-kktcm" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.403973 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrjwv\" (UniqueName: \"kubernetes.io/projected/b1d3e297-8696-43ae-84f8-da003e5a03e5-kube-api-access-mrjwv\") pod \"nova-cell0-db-create-tp852\" (UID: \"b1d3e297-8696-43ae-84f8-da003e5a03e5\") " pod="openstack/nova-cell0-db-create-tp852" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.404038 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79afaafd-5f51-4862-9c09-e7f817d54f73-operator-scripts\") pod \"nova-cell1-db-create-kktcm\" (UID: \"79afaafd-5f51-4862-9c09-e7f817d54f73\") " pod="openstack/nova-cell1-db-create-kktcm" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.404066 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1d3e297-8696-43ae-84f8-da003e5a03e5-operator-scripts\") pod \"nova-cell0-db-create-tp852\" (UID: \"b1d3e297-8696-43ae-84f8-da003e5a03e5\") " pod="openstack/nova-cell0-db-create-tp852" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.404897 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1d3e297-8696-43ae-84f8-da003e5a03e5-operator-scripts\") pod \"nova-cell0-db-create-tp852\" (UID: \"b1d3e297-8696-43ae-84f8-da003e5a03e5\") " pod="openstack/nova-cell0-db-create-tp852" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.457603 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrjwv\" (UniqueName: \"kubernetes.io/projected/b1d3e297-8696-43ae-84f8-da003e5a03e5-kube-api-access-mrjwv\") pod \"nova-cell0-db-create-tp852\" (UID: \"b1d3e297-8696-43ae-84f8-da003e5a03e5\") " pod="openstack/nova-cell0-db-create-tp852" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.493121 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-t898j" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.508561 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79afaafd-5f51-4862-9c09-e7f817d54f73-operator-scripts\") pod \"nova-cell1-db-create-kktcm\" (UID: \"79afaafd-5f51-4862-9c09-e7f817d54f73\") " pod="openstack/nova-cell1-db-create-kktcm" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.508655 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztqns\" (UniqueName: \"kubernetes.io/projected/79afaafd-5f51-4862-9c09-e7f817d54f73-kube-api-access-ztqns\") pod \"nova-cell1-db-create-kktcm\" (UID: \"79afaafd-5f51-4862-9c09-e7f817d54f73\") " pod="openstack/nova-cell1-db-create-kktcm" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.514108 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79afaafd-5f51-4862-9c09-e7f817d54f73-operator-scripts\") pod \"nova-cell1-db-create-kktcm\" (UID: \"79afaafd-5f51-4862-9c09-e7f817d54f73\") " pod="openstack/nova-cell1-db-create-kktcm" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.521016 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-d2d8-account-create-tfkjf"] Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.549604 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.553888 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztqns\" (UniqueName: \"kubernetes.io/projected/79afaafd-5f51-4862-9c09-e7f817d54f73-kube-api-access-ztqns\") pod \"nova-cell1-db-create-kktcm\" (UID: \"79afaafd-5f51-4862-9c09-e7f817d54f73\") " pod="openstack/nova-cell1-db-create-kktcm" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.616554 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-tp852" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.618426 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78fcad7c-4049-4dae-9fa1-2f8353447e12-operator-scripts\") pod \"nova-api-d2d8-account-create-tfkjf\" (UID: \"78fcad7c-4049-4dae-9fa1-2f8353447e12\") " pod="openstack/nova-api-d2d8-account-create-tfkjf" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.623981 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92rkt\" (UniqueName: \"kubernetes.io/projected/78fcad7c-4049-4dae-9fa1-2f8353447e12-kube-api-access-92rkt\") pod \"nova-api-d2d8-account-create-tfkjf\" (UID: \"78fcad7c-4049-4dae-9fa1-2f8353447e12\") " pod="openstack/nova-api-d2d8-account-create-tfkjf" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.657372 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-ade7-account-create-mj2zg"] Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.659809 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ade7-account-create-mj2zg" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.663135 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-ade7-account-create-mj2zg"] Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.663477 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.732969 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/096bd741-6394-4f58-9024-d425d96a0bd7-operator-scripts\") pod \"nova-cell0-ade7-account-create-mj2zg\" (UID: \"096bd741-6394-4f58-9024-d425d96a0bd7\") " pod="openstack/nova-cell0-ade7-account-create-mj2zg" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.733026 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92rkt\" (UniqueName: \"kubernetes.io/projected/78fcad7c-4049-4dae-9fa1-2f8353447e12-kube-api-access-92rkt\") pod \"nova-api-d2d8-account-create-tfkjf\" (UID: \"78fcad7c-4049-4dae-9fa1-2f8353447e12\") " pod="openstack/nova-api-d2d8-account-create-tfkjf" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.733285 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4zxb\" (UniqueName: \"kubernetes.io/projected/096bd741-6394-4f58-9024-d425d96a0bd7-kube-api-access-b4zxb\") pod \"nova-cell0-ade7-account-create-mj2zg\" (UID: \"096bd741-6394-4f58-9024-d425d96a0bd7\") " pod="openstack/nova-cell0-ade7-account-create-mj2zg" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.733326 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78fcad7c-4049-4dae-9fa1-2f8353447e12-operator-scripts\") pod \"nova-api-d2d8-account-create-tfkjf\" (UID: \"78fcad7c-4049-4dae-9fa1-2f8353447e12\") " pod="openstack/nova-api-d2d8-account-create-tfkjf" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.734419 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78fcad7c-4049-4dae-9fa1-2f8353447e12-operator-scripts\") pod \"nova-api-d2d8-account-create-tfkjf\" (UID: \"78fcad7c-4049-4dae-9fa1-2f8353447e12\") " pod="openstack/nova-api-d2d8-account-create-tfkjf" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.758325 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92rkt\" (UniqueName: \"kubernetes.io/projected/78fcad7c-4049-4dae-9fa1-2f8353447e12-kube-api-access-92rkt\") pod \"nova-api-d2d8-account-create-tfkjf\" (UID: \"78fcad7c-4049-4dae-9fa1-2f8353447e12\") " pod="openstack/nova-api-d2d8-account-create-tfkjf" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.779196 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-974e-account-create-qw2ds"] Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.780922 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-974e-account-create-qw2ds" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.790898 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.791476 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kktcm" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.842000 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4zxb\" (UniqueName: \"kubernetes.io/projected/096bd741-6394-4f58-9024-d425d96a0bd7-kube-api-access-b4zxb\") pod \"nova-cell0-ade7-account-create-mj2zg\" (UID: \"096bd741-6394-4f58-9024-d425d96a0bd7\") " pod="openstack/nova-cell0-ade7-account-create-mj2zg" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.842110 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/096bd741-6394-4f58-9024-d425d96a0bd7-operator-scripts\") pod \"nova-cell0-ade7-account-create-mj2zg\" (UID: \"096bd741-6394-4f58-9024-d425d96a0bd7\") " pod="openstack/nova-cell0-ade7-account-create-mj2zg" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.842199 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4-operator-scripts\") pod \"nova-cell1-974e-account-create-qw2ds\" (UID: \"fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4\") " pod="openstack/nova-cell1-974e-account-create-qw2ds" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.842253 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhclk\" (UniqueName: \"kubernetes.io/projected/fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4-kube-api-access-jhclk\") pod \"nova-cell1-974e-account-create-qw2ds\" (UID: \"fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4\") " pod="openstack/nova-cell1-974e-account-create-qw2ds" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.843872 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-974e-account-create-qw2ds"] Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.849880 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/096bd741-6394-4f58-9024-d425d96a0bd7-operator-scripts\") pod \"nova-cell0-ade7-account-create-mj2zg\" (UID: \"096bd741-6394-4f58-9024-d425d96a0bd7\") " pod="openstack/nova-cell0-ade7-account-create-mj2zg" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.861160 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4zxb\" (UniqueName: \"kubernetes.io/projected/096bd741-6394-4f58-9024-d425d96a0bd7-kube-api-access-b4zxb\") pod \"nova-cell0-ade7-account-create-mj2zg\" (UID: \"096bd741-6394-4f58-9024-d425d96a0bd7\") " pod="openstack/nova-cell0-ade7-account-create-mj2zg" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.941489 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-d2d8-account-create-tfkjf" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.946968 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4-operator-scripts\") pod \"nova-cell1-974e-account-create-qw2ds\" (UID: \"fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4\") " pod="openstack/nova-cell1-974e-account-create-qw2ds" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.947053 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhclk\" (UniqueName: \"kubernetes.io/projected/fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4-kube-api-access-jhclk\") pod \"nova-cell1-974e-account-create-qw2ds\" (UID: \"fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4\") " pod="openstack/nova-cell1-974e-account-create-qw2ds" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.948148 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4-operator-scripts\") pod \"nova-cell1-974e-account-create-qw2ds\" (UID: \"fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4\") " pod="openstack/nova-cell1-974e-account-create-qw2ds" Nov 23 07:00:31 crc kubenswrapper[4559]: I1123 07:00:31.979175 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhclk\" (UniqueName: \"kubernetes.io/projected/fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4-kube-api-access-jhclk\") pod \"nova-cell1-974e-account-create-qw2ds\" (UID: \"fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4\") " pod="openstack/nova-cell1-974e-account-create-qw2ds" Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.043844 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ade7-account-create-mj2zg" Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.087962 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-t898j"] Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.145287 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-974e-account-create-qw2ds" Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.164161 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-tp852"] Nov 23 07:00:32 crc kubenswrapper[4559]: W1123 07:00:32.226764 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1d3e297_8696_43ae_84f8_da003e5a03e5.slice/crio-a5b9fac63e0f0e877b763669fc0cec4cf4950d70a1fd7947641d7d13e9260d82 WatchSource:0}: Error finding container a5b9fac63e0f0e877b763669fc0cec4cf4950d70a1fd7947641d7d13e9260d82: Status 404 returned error can't find the container with id a5b9fac63e0f0e877b763669fc0cec4cf4950d70a1fd7947641d7d13e9260d82 Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.358895 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-kktcm"] Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.472346 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-d2d8-account-create-tfkjf"] Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.537964 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.541259 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-ade7-account-create-mj2zg"] Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.556021 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b0ee294-003c-49b1-878e-50097454795f-scripts\") pod \"5b0ee294-003c-49b1-878e-50097454795f\" (UID: \"5b0ee294-003c-49b1-878e-50097454795f\") " Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.556062 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b0ee294-003c-49b1-878e-50097454795f-run-httpd\") pod \"5b0ee294-003c-49b1-878e-50097454795f\" (UID: \"5b0ee294-003c-49b1-878e-50097454795f\") " Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.556130 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-824mc\" (UniqueName: \"kubernetes.io/projected/5b0ee294-003c-49b1-878e-50097454795f-kube-api-access-824mc\") pod \"5b0ee294-003c-49b1-878e-50097454795f\" (UID: \"5b0ee294-003c-49b1-878e-50097454795f\") " Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.556180 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b0ee294-003c-49b1-878e-50097454795f-combined-ca-bundle\") pod \"5b0ee294-003c-49b1-878e-50097454795f\" (UID: \"5b0ee294-003c-49b1-878e-50097454795f\") " Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.556221 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b0ee294-003c-49b1-878e-50097454795f-sg-core-conf-yaml\") pod \"5b0ee294-003c-49b1-878e-50097454795f\" (UID: \"5b0ee294-003c-49b1-878e-50097454795f\") " Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.556249 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b0ee294-003c-49b1-878e-50097454795f-config-data\") pod \"5b0ee294-003c-49b1-878e-50097454795f\" (UID: \"5b0ee294-003c-49b1-878e-50097454795f\") " Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.556277 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b0ee294-003c-49b1-878e-50097454795f-log-httpd\") pod \"5b0ee294-003c-49b1-878e-50097454795f\" (UID: \"5b0ee294-003c-49b1-878e-50097454795f\") " Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.557066 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b0ee294-003c-49b1-878e-50097454795f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5b0ee294-003c-49b1-878e-50097454795f" (UID: "5b0ee294-003c-49b1-878e-50097454795f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.558042 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b0ee294-003c-49b1-878e-50097454795f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5b0ee294-003c-49b1-878e-50097454795f" (UID: "5b0ee294-003c-49b1-878e-50097454795f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.572950 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-d2d8-account-create-tfkjf" event={"ID":"78fcad7c-4049-4dae-9fa1-2f8353447e12","Type":"ContainerStarted","Data":"d982566ecaf1d1196920ea57c167bb5c94062758e567d86d78eb32d05b68029f"} Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.577753 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-kktcm" event={"ID":"79afaafd-5f51-4862-9c09-e7f817d54f73","Type":"ContainerStarted","Data":"b68662a211784edef9e35741c2a860eb3184e5a9ee5cce5b7e462ebc279c4622"} Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.590415 4559 generic.go:334] "Generic (PLEG): container finished" podID="5b0ee294-003c-49b1-878e-50097454795f" containerID="46666f0629352c2d81b02e541670d6640a217e0d5b21a838ed83703abd0d91ef" exitCode=0 Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.590511 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.590529 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b0ee294-003c-49b1-878e-50097454795f","Type":"ContainerDied","Data":"46666f0629352c2d81b02e541670d6640a217e0d5b21a838ed83703abd0d91ef"} Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.591262 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b0ee294-003c-49b1-878e-50097454795f","Type":"ContainerDied","Data":"9381f2fde154e6c805142313edda953a39caa0a3264e48406db0f56b6604ad69"} Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.591318 4559 scope.go:117] "RemoveContainer" containerID="81ccb4b66519f8f7d2facac1d0226906cf29314b35e5f091f2c049fa273ef2e2" Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.600525 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b0ee294-003c-49b1-878e-50097454795f-scripts" (OuterVolumeSpecName: "scripts") pod "5b0ee294-003c-49b1-878e-50097454795f" (UID: "5b0ee294-003c-49b1-878e-50097454795f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.600685 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-t898j" event={"ID":"ec53c68f-1637-47e8-8941-0ada991639ec","Type":"ContainerStarted","Data":"6f027cc235e87ebc22c09ef418681e2c53a6dad2c2eec69ac930bbd1af843d58"} Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.601163 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-t898j" event={"ID":"ec53c68f-1637-47e8-8941-0ada991639ec","Type":"ContainerStarted","Data":"c4153b9d113e617e989c9c9cbbbcc00834cef21c275435c3ca5a77b53fb4700d"} Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.603055 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-tp852" event={"ID":"b1d3e297-8696-43ae-84f8-da003e5a03e5","Type":"ContainerStarted","Data":"c23d564932376ae305fa54fa32a91e2bae72476f83833fbf5561e5956d153e8f"} Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.603480 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-tp852" event={"ID":"b1d3e297-8696-43ae-84f8-da003e5a03e5","Type":"ContainerStarted","Data":"a5b9fac63e0f0e877b763669fc0cec4cf4950d70a1fd7947641d7d13e9260d82"} Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.612020 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b0ee294-003c-49b1-878e-50097454795f-kube-api-access-824mc" (OuterVolumeSpecName: "kube-api-access-824mc") pod "5b0ee294-003c-49b1-878e-50097454795f" (UID: "5b0ee294-003c-49b1-878e-50097454795f"). InnerVolumeSpecName "kube-api-access-824mc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.660085 4559 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b0ee294-003c-49b1-878e-50097454795f-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.660112 4559 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b0ee294-003c-49b1-878e-50097454795f-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.660125 4559 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b0ee294-003c-49b1-878e-50097454795f-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.660134 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-824mc\" (UniqueName: \"kubernetes.io/projected/5b0ee294-003c-49b1-878e-50097454795f-kube-api-access-824mc\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.663335 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b0ee294-003c-49b1-878e-50097454795f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5b0ee294-003c-49b1-878e-50097454795f" (UID: "5b0ee294-003c-49b1-878e-50097454795f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.697186 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-974e-account-create-qw2ds"] Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.709556 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b0ee294-003c-49b1-878e-50097454795f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b0ee294-003c-49b1-878e-50097454795f" (UID: "5b0ee294-003c-49b1-878e-50097454795f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.751385 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b0ee294-003c-49b1-878e-50097454795f-config-data" (OuterVolumeSpecName: "config-data") pod "5b0ee294-003c-49b1-878e-50097454795f" (UID: "5b0ee294-003c-49b1-878e-50097454795f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.762080 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b0ee294-003c-49b1-878e-50097454795f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.762105 4559 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b0ee294-003c-49b1-878e-50097454795f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:32 crc kubenswrapper[4559]: I1123 07:00:32.762115 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b0ee294-003c-49b1-878e-50097454795f-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.014900 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.039991 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.050013 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:00:33 crc kubenswrapper[4559]: E1123 07:00:33.050437 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b0ee294-003c-49b1-878e-50097454795f" containerName="sg-core" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.050453 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b0ee294-003c-49b1-878e-50097454795f" containerName="sg-core" Nov 23 07:00:33 crc kubenswrapper[4559]: E1123 07:00:33.050499 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b0ee294-003c-49b1-878e-50097454795f" containerName="proxy-httpd" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.050504 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b0ee294-003c-49b1-878e-50097454795f" containerName="proxy-httpd" Nov 23 07:00:33 crc kubenswrapper[4559]: E1123 07:00:33.050513 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b0ee294-003c-49b1-878e-50097454795f" containerName="ceilometer-central-agent" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.050519 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b0ee294-003c-49b1-878e-50097454795f" containerName="ceilometer-central-agent" Nov 23 07:00:33 crc kubenswrapper[4559]: E1123 07:00:33.050525 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b0ee294-003c-49b1-878e-50097454795f" containerName="ceilometer-notification-agent" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.050532 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b0ee294-003c-49b1-878e-50097454795f" containerName="ceilometer-notification-agent" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.050712 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b0ee294-003c-49b1-878e-50097454795f" containerName="ceilometer-central-agent" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.050732 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b0ee294-003c-49b1-878e-50097454795f" containerName="ceilometer-notification-agent" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.050741 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b0ee294-003c-49b1-878e-50097454795f" containerName="sg-core" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.050753 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b0ee294-003c-49b1-878e-50097454795f" containerName="proxy-httpd" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.052286 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.055603 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.055791 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.060174 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.173167 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c3cf2c6-0694-425f-86e0-5313595a152a-log-httpd\") pod \"ceilometer-0\" (UID: \"9c3cf2c6-0694-425f-86e0-5313595a152a\") " pod="openstack/ceilometer-0" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.173220 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c3cf2c6-0694-425f-86e0-5313595a152a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9c3cf2c6-0694-425f-86e0-5313595a152a\") " pod="openstack/ceilometer-0" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.173437 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qws98\" (UniqueName: \"kubernetes.io/projected/9c3cf2c6-0694-425f-86e0-5313595a152a-kube-api-access-qws98\") pod \"ceilometer-0\" (UID: \"9c3cf2c6-0694-425f-86e0-5313595a152a\") " pod="openstack/ceilometer-0" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.173546 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c3cf2c6-0694-425f-86e0-5313595a152a-scripts\") pod \"ceilometer-0\" (UID: \"9c3cf2c6-0694-425f-86e0-5313595a152a\") " pod="openstack/ceilometer-0" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.173701 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c3cf2c6-0694-425f-86e0-5313595a152a-run-httpd\") pod \"ceilometer-0\" (UID: \"9c3cf2c6-0694-425f-86e0-5313595a152a\") " pod="openstack/ceilometer-0" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.173796 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9c3cf2c6-0694-425f-86e0-5313595a152a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9c3cf2c6-0694-425f-86e0-5313595a152a\") " pod="openstack/ceilometer-0" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.173981 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c3cf2c6-0694-425f-86e0-5313595a152a-config-data\") pod \"ceilometer-0\" (UID: \"9c3cf2c6-0694-425f-86e0-5313595a152a\") " pod="openstack/ceilometer-0" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.275555 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c3cf2c6-0694-425f-86e0-5313595a152a-run-httpd\") pod \"ceilometer-0\" (UID: \"9c3cf2c6-0694-425f-86e0-5313595a152a\") " pod="openstack/ceilometer-0" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.275606 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9c3cf2c6-0694-425f-86e0-5313595a152a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9c3cf2c6-0694-425f-86e0-5313595a152a\") " pod="openstack/ceilometer-0" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.275729 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c3cf2c6-0694-425f-86e0-5313595a152a-config-data\") pod \"ceilometer-0\" (UID: \"9c3cf2c6-0694-425f-86e0-5313595a152a\") " pod="openstack/ceilometer-0" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.275785 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c3cf2c6-0694-425f-86e0-5313595a152a-log-httpd\") pod \"ceilometer-0\" (UID: \"9c3cf2c6-0694-425f-86e0-5313595a152a\") " pod="openstack/ceilometer-0" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.275813 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c3cf2c6-0694-425f-86e0-5313595a152a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9c3cf2c6-0694-425f-86e0-5313595a152a\") " pod="openstack/ceilometer-0" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.275859 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qws98\" (UniqueName: \"kubernetes.io/projected/9c3cf2c6-0694-425f-86e0-5313595a152a-kube-api-access-qws98\") pod \"ceilometer-0\" (UID: \"9c3cf2c6-0694-425f-86e0-5313595a152a\") " pod="openstack/ceilometer-0" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.275891 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c3cf2c6-0694-425f-86e0-5313595a152a-scripts\") pod \"ceilometer-0\" (UID: \"9c3cf2c6-0694-425f-86e0-5313595a152a\") " pod="openstack/ceilometer-0" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.275990 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c3cf2c6-0694-425f-86e0-5313595a152a-run-httpd\") pod \"ceilometer-0\" (UID: \"9c3cf2c6-0694-425f-86e0-5313595a152a\") " pod="openstack/ceilometer-0" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.276357 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c3cf2c6-0694-425f-86e0-5313595a152a-log-httpd\") pod \"ceilometer-0\" (UID: \"9c3cf2c6-0694-425f-86e0-5313595a152a\") " pod="openstack/ceilometer-0" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.280676 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c3cf2c6-0694-425f-86e0-5313595a152a-scripts\") pod \"ceilometer-0\" (UID: \"9c3cf2c6-0694-425f-86e0-5313595a152a\") " pod="openstack/ceilometer-0" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.280828 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9c3cf2c6-0694-425f-86e0-5313595a152a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9c3cf2c6-0694-425f-86e0-5313595a152a\") " pod="openstack/ceilometer-0" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.281038 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c3cf2c6-0694-425f-86e0-5313595a152a-config-data\") pod \"ceilometer-0\" (UID: \"9c3cf2c6-0694-425f-86e0-5313595a152a\") " pod="openstack/ceilometer-0" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.281919 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c3cf2c6-0694-425f-86e0-5313595a152a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9c3cf2c6-0694-425f-86e0-5313595a152a\") " pod="openstack/ceilometer-0" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.295870 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qws98\" (UniqueName: \"kubernetes.io/projected/9c3cf2c6-0694-425f-86e0-5313595a152a-kube-api-access-qws98\") pod \"ceilometer-0\" (UID: \"9c3cf2c6-0694-425f-86e0-5313595a152a\") " pod="openstack/ceilometer-0" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.376224 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.616887 4559 generic.go:334] "Generic (PLEG): container finished" podID="ec53c68f-1637-47e8-8941-0ada991639ec" containerID="6f027cc235e87ebc22c09ef418681e2c53a6dad2c2eec69ac930bbd1af843d58" exitCode=0 Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.616989 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-t898j" event={"ID":"ec53c68f-1637-47e8-8941-0ada991639ec","Type":"ContainerDied","Data":"6f027cc235e87ebc22c09ef418681e2c53a6dad2c2eec69ac930bbd1af843d58"} Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.631276 4559 generic.go:334] "Generic (PLEG): container finished" podID="b1d3e297-8696-43ae-84f8-da003e5a03e5" containerID="c23d564932376ae305fa54fa32a91e2bae72476f83833fbf5561e5956d153e8f" exitCode=0 Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.632078 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-tp852" event={"ID":"b1d3e297-8696-43ae-84f8-da003e5a03e5","Type":"ContainerDied","Data":"c23d564932376ae305fa54fa32a91e2bae72476f83833fbf5561e5956d153e8f"} Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.635176 4559 generic.go:334] "Generic (PLEG): container finished" podID="fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4" containerID="e626d8fb33c8d679398bf1d7cdc6807e0533a46669e270a980cda59efc314eb1" exitCode=0 Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.635248 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-974e-account-create-qw2ds" event={"ID":"fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4","Type":"ContainerDied","Data":"e626d8fb33c8d679398bf1d7cdc6807e0533a46669e270a980cda59efc314eb1"} Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.635277 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-974e-account-create-qw2ds" event={"ID":"fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4","Type":"ContainerStarted","Data":"8f031e2f3a5c14d534dd7b4ae0fb33c282d6f6ff97298f31bb734f9e3f7ac9a2"} Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.639930 4559 generic.go:334] "Generic (PLEG): container finished" podID="78fcad7c-4049-4dae-9fa1-2f8353447e12" containerID="5c1643a15abf22c3a043c168790f7e71a2c2d53736e6df77cc37dfea667c43a7" exitCode=0 Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.640011 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-d2d8-account-create-tfkjf" event={"ID":"78fcad7c-4049-4dae-9fa1-2f8353447e12","Type":"ContainerDied","Data":"5c1643a15abf22c3a043c168790f7e71a2c2d53736e6df77cc37dfea667c43a7"} Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.642190 4559 generic.go:334] "Generic (PLEG): container finished" podID="79afaafd-5f51-4862-9c09-e7f817d54f73" containerID="25386b49570cd3a8fc6b9401d205700597ac4704ba860466f5ad40289cf95742" exitCode=0 Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.642246 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-kktcm" event={"ID":"79afaafd-5f51-4862-9c09-e7f817d54f73","Type":"ContainerDied","Data":"25386b49570cd3a8fc6b9401d205700597ac4704ba860466f5ad40289cf95742"} Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.645726 4559 generic.go:334] "Generic (PLEG): container finished" podID="096bd741-6394-4f58-9024-d425d96a0bd7" containerID="8a0ee1424748e596b9fdc04cc74255047de5f668af4056f42c70810a28a7af69" exitCode=0 Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.645761 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ade7-account-create-mj2zg" event={"ID":"096bd741-6394-4f58-9024-d425d96a0bd7","Type":"ContainerDied","Data":"8a0ee1424748e596b9fdc04cc74255047de5f668af4056f42c70810a28a7af69"} Nov 23 07:00:33 crc kubenswrapper[4559]: I1123 07:00:33.645791 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ade7-account-create-mj2zg" event={"ID":"096bd741-6394-4f58-9024-d425d96a0bd7","Type":"ContainerStarted","Data":"c6c08850f922460ff6b39daa302fdb1bc6c430a0cf0324bb5b56e95a27442a08"} Nov 23 07:00:34 crc kubenswrapper[4559]: I1123 07:00:34.300317 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b0ee294-003c-49b1-878e-50097454795f" path="/var/lib/kubelet/pods/5b0ee294-003c-49b1-878e-50097454795f/volumes" Nov 23 07:00:35 crc kubenswrapper[4559]: I1123 07:00:35.267446 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 07:00:35 crc kubenswrapper[4559]: I1123 07:00:35.268410 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="c3a4ae67-888b-46c2-bad1-92758aa2ea4f" containerName="glance-log" containerID="cri-o://a87a2cb07397e7a692e575ea7ebc473172586952ba546154c48fc11d0195ff43" gracePeriod=30 Nov 23 07:00:35 crc kubenswrapper[4559]: I1123 07:00:35.268497 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="c3a4ae67-888b-46c2-bad1-92758aa2ea4f" containerName="glance-httpd" containerID="cri-o://27f8bbb0ad0a072b9c09d2f2502d82cb90133957a9ee3c475747e68d2e5c9159" gracePeriod=30 Nov 23 07:00:35 crc kubenswrapper[4559]: I1123 07:00:35.674961 4559 generic.go:334] "Generic (PLEG): container finished" podID="c3a4ae67-888b-46c2-bad1-92758aa2ea4f" containerID="a87a2cb07397e7a692e575ea7ebc473172586952ba546154c48fc11d0195ff43" exitCode=143 Nov 23 07:00:35 crc kubenswrapper[4559]: I1123 07:00:35.675005 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c3a4ae67-888b-46c2-bad1-92758aa2ea4f","Type":"ContainerDied","Data":"a87a2cb07397e7a692e575ea7ebc473172586952ba546154c48fc11d0195ff43"} Nov 23 07:00:37 crc kubenswrapper[4559]: I1123 07:00:37.090345 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 07:00:37 crc kubenswrapper[4559]: I1123 07:00:37.091360 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="bf03ed6e-e623-4ee8-afd4-361d6d18097a" containerName="glance-log" containerID="cri-o://a2503ed47292a8e094ec1cee0b8416e38dfefebe40c846cff1fd8c7958096c24" gracePeriod=30 Nov 23 07:00:37 crc kubenswrapper[4559]: I1123 07:00:37.091689 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="bf03ed6e-e623-4ee8-afd4-361d6d18097a" containerName="glance-httpd" containerID="cri-o://cbd4d0e094e31047191103ed3e0e5d8a94e077ebe5cb31c5ec25c3b7850932a0" gracePeriod=30 Nov 23 07:00:37 crc kubenswrapper[4559]: I1123 07:00:37.695605 4559 generic.go:334] "Generic (PLEG): container finished" podID="bf03ed6e-e623-4ee8-afd4-361d6d18097a" containerID="a2503ed47292a8e094ec1cee0b8416e38dfefebe40c846cff1fd8c7958096c24" exitCode=143 Nov 23 07:00:37 crc kubenswrapper[4559]: I1123 07:00:37.695803 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bf03ed6e-e623-4ee8-afd4-361d6d18097a","Type":"ContainerDied","Data":"a2503ed47292a8e094ec1cee0b8416e38dfefebe40c846cff1fd8c7958096c24"} Nov 23 07:00:37 crc kubenswrapper[4559]: I1123 07:00:37.816444 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:00:37 crc kubenswrapper[4559]: I1123 07:00:37.842796 4559 scope.go:117] "RemoveContainer" containerID="ae1b972ec5cc95c188cb0fac2c7e0ec994b95f2fa2655a84967d824390d1ceb4" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.051896 4559 scope.go:117] "RemoveContainer" containerID="46666f0629352c2d81b02e541670d6640a217e0d5b21a838ed83703abd0d91ef" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.063609 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kktcm" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.067241 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-t898j" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.087560 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-d2d8-account-create-tfkjf" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.094872 4559 scope.go:117] "RemoveContainer" containerID="c21b860c45243f8824cd1b77c0a54694e67b445eac3577a219d0e5d4f7e8ce15" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.096623 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ade7-account-create-mj2zg" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.096655 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-tp852" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.128630 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-974e-account-create-qw2ds" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.140605 4559 scope.go:117] "RemoveContainer" containerID="81ccb4b66519f8f7d2facac1d0226906cf29314b35e5f091f2c049fa273ef2e2" Nov 23 07:00:38 crc kubenswrapper[4559]: E1123 07:00:38.142780 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81ccb4b66519f8f7d2facac1d0226906cf29314b35e5f091f2c049fa273ef2e2\": container with ID starting with 81ccb4b66519f8f7d2facac1d0226906cf29314b35e5f091f2c049fa273ef2e2 not found: ID does not exist" containerID="81ccb4b66519f8f7d2facac1d0226906cf29314b35e5f091f2c049fa273ef2e2" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.142847 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81ccb4b66519f8f7d2facac1d0226906cf29314b35e5f091f2c049fa273ef2e2"} err="failed to get container status \"81ccb4b66519f8f7d2facac1d0226906cf29314b35e5f091f2c049fa273ef2e2\": rpc error: code = NotFound desc = could not find container \"81ccb4b66519f8f7d2facac1d0226906cf29314b35e5f091f2c049fa273ef2e2\": container with ID starting with 81ccb4b66519f8f7d2facac1d0226906cf29314b35e5f091f2c049fa273ef2e2 not found: ID does not exist" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.142877 4559 scope.go:117] "RemoveContainer" containerID="ae1b972ec5cc95c188cb0fac2c7e0ec994b95f2fa2655a84967d824390d1ceb4" Nov 23 07:00:38 crc kubenswrapper[4559]: E1123 07:00:38.143207 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae1b972ec5cc95c188cb0fac2c7e0ec994b95f2fa2655a84967d824390d1ceb4\": container with ID starting with ae1b972ec5cc95c188cb0fac2c7e0ec994b95f2fa2655a84967d824390d1ceb4 not found: ID does not exist" containerID="ae1b972ec5cc95c188cb0fac2c7e0ec994b95f2fa2655a84967d824390d1ceb4" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.143254 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae1b972ec5cc95c188cb0fac2c7e0ec994b95f2fa2655a84967d824390d1ceb4"} err="failed to get container status \"ae1b972ec5cc95c188cb0fac2c7e0ec994b95f2fa2655a84967d824390d1ceb4\": rpc error: code = NotFound desc = could not find container \"ae1b972ec5cc95c188cb0fac2c7e0ec994b95f2fa2655a84967d824390d1ceb4\": container with ID starting with ae1b972ec5cc95c188cb0fac2c7e0ec994b95f2fa2655a84967d824390d1ceb4 not found: ID does not exist" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.143270 4559 scope.go:117] "RemoveContainer" containerID="46666f0629352c2d81b02e541670d6640a217e0d5b21a838ed83703abd0d91ef" Nov 23 07:00:38 crc kubenswrapper[4559]: E1123 07:00:38.143939 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46666f0629352c2d81b02e541670d6640a217e0d5b21a838ed83703abd0d91ef\": container with ID starting with 46666f0629352c2d81b02e541670d6640a217e0d5b21a838ed83703abd0d91ef not found: ID does not exist" containerID="46666f0629352c2d81b02e541670d6640a217e0d5b21a838ed83703abd0d91ef" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.144004 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46666f0629352c2d81b02e541670d6640a217e0d5b21a838ed83703abd0d91ef"} err="failed to get container status \"46666f0629352c2d81b02e541670d6640a217e0d5b21a838ed83703abd0d91ef\": rpc error: code = NotFound desc = could not find container \"46666f0629352c2d81b02e541670d6640a217e0d5b21a838ed83703abd0d91ef\": container with ID starting with 46666f0629352c2d81b02e541670d6640a217e0d5b21a838ed83703abd0d91ef not found: ID does not exist" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.144021 4559 scope.go:117] "RemoveContainer" containerID="c21b860c45243f8824cd1b77c0a54694e67b445eac3577a219d0e5d4f7e8ce15" Nov 23 07:00:38 crc kubenswrapper[4559]: E1123 07:00:38.145444 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c21b860c45243f8824cd1b77c0a54694e67b445eac3577a219d0e5d4f7e8ce15\": container with ID starting with c21b860c45243f8824cd1b77c0a54694e67b445eac3577a219d0e5d4f7e8ce15 not found: ID does not exist" containerID="c21b860c45243f8824cd1b77c0a54694e67b445eac3577a219d0e5d4f7e8ce15" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.145468 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c21b860c45243f8824cd1b77c0a54694e67b445eac3577a219d0e5d4f7e8ce15"} err="failed to get container status \"c21b860c45243f8824cd1b77c0a54694e67b445eac3577a219d0e5d4f7e8ce15\": rpc error: code = NotFound desc = could not find container \"c21b860c45243f8824cd1b77c0a54694e67b445eac3577a219d0e5d4f7e8ce15\": container with ID starting with c21b860c45243f8824cd1b77c0a54694e67b445eac3577a219d0e5d4f7e8ce15 not found: ID does not exist" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.191536 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78fcad7c-4049-4dae-9fa1-2f8353447e12-operator-scripts\") pod \"78fcad7c-4049-4dae-9fa1-2f8353447e12\" (UID: \"78fcad7c-4049-4dae-9fa1-2f8353447e12\") " Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.191744 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79afaafd-5f51-4862-9c09-e7f817d54f73-operator-scripts\") pod \"79afaafd-5f51-4862-9c09-e7f817d54f73\" (UID: \"79afaafd-5f51-4862-9c09-e7f817d54f73\") " Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.191797 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec53c68f-1637-47e8-8941-0ada991639ec-operator-scripts\") pod \"ec53c68f-1637-47e8-8941-0ada991639ec\" (UID: \"ec53c68f-1637-47e8-8941-0ada991639ec\") " Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.191916 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stqxc\" (UniqueName: \"kubernetes.io/projected/ec53c68f-1637-47e8-8941-0ada991639ec-kube-api-access-stqxc\") pod \"ec53c68f-1637-47e8-8941-0ada991639ec\" (UID: \"ec53c68f-1637-47e8-8941-0ada991639ec\") " Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.191954 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztqns\" (UniqueName: \"kubernetes.io/projected/79afaafd-5f51-4862-9c09-e7f817d54f73-kube-api-access-ztqns\") pod \"79afaafd-5f51-4862-9c09-e7f817d54f73\" (UID: \"79afaafd-5f51-4862-9c09-e7f817d54f73\") " Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.191992 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92rkt\" (UniqueName: \"kubernetes.io/projected/78fcad7c-4049-4dae-9fa1-2f8353447e12-kube-api-access-92rkt\") pod \"78fcad7c-4049-4dae-9fa1-2f8353447e12\" (UID: \"78fcad7c-4049-4dae-9fa1-2f8353447e12\") " Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.192353 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78fcad7c-4049-4dae-9fa1-2f8353447e12-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "78fcad7c-4049-4dae-9fa1-2f8353447e12" (UID: "78fcad7c-4049-4dae-9fa1-2f8353447e12"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.192461 4559 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78fcad7c-4049-4dae-9fa1-2f8353447e12-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.193552 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec53c68f-1637-47e8-8941-0ada991639ec-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ec53c68f-1637-47e8-8941-0ada991639ec" (UID: "ec53c68f-1637-47e8-8941-0ada991639ec"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.193561 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79afaafd-5f51-4862-9c09-e7f817d54f73-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "79afaafd-5f51-4862-9c09-e7f817d54f73" (UID: "79afaafd-5f51-4862-9c09-e7f817d54f73"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.196974 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79afaafd-5f51-4862-9c09-e7f817d54f73-kube-api-access-ztqns" (OuterVolumeSpecName: "kube-api-access-ztqns") pod "79afaafd-5f51-4862-9c09-e7f817d54f73" (UID: "79afaafd-5f51-4862-9c09-e7f817d54f73"). InnerVolumeSpecName "kube-api-access-ztqns". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.197556 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78fcad7c-4049-4dae-9fa1-2f8353447e12-kube-api-access-92rkt" (OuterVolumeSpecName: "kube-api-access-92rkt") pod "78fcad7c-4049-4dae-9fa1-2f8353447e12" (UID: "78fcad7c-4049-4dae-9fa1-2f8353447e12"). InnerVolumeSpecName "kube-api-access-92rkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.197613 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec53c68f-1637-47e8-8941-0ada991639ec-kube-api-access-stqxc" (OuterVolumeSpecName: "kube-api-access-stqxc") pod "ec53c68f-1637-47e8-8941-0ada991639ec" (UID: "ec53c68f-1637-47e8-8941-0ada991639ec"). InnerVolumeSpecName "kube-api-access-stqxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.293110 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4-operator-scripts\") pod \"fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4\" (UID: \"fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4\") " Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.293322 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/096bd741-6394-4f58-9024-d425d96a0bd7-operator-scripts\") pod \"096bd741-6394-4f58-9024-d425d96a0bd7\" (UID: \"096bd741-6394-4f58-9024-d425d96a0bd7\") " Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.293360 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4zxb\" (UniqueName: \"kubernetes.io/projected/096bd741-6394-4f58-9024-d425d96a0bd7-kube-api-access-b4zxb\") pod \"096bd741-6394-4f58-9024-d425d96a0bd7\" (UID: \"096bd741-6394-4f58-9024-d425d96a0bd7\") " Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.293412 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhclk\" (UniqueName: \"kubernetes.io/projected/fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4-kube-api-access-jhclk\") pod \"fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4\" (UID: \"fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4\") " Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.293517 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4" (UID: "fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.293663 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1d3e297-8696-43ae-84f8-da003e5a03e5-operator-scripts\") pod \"b1d3e297-8696-43ae-84f8-da003e5a03e5\" (UID: \"b1d3e297-8696-43ae-84f8-da003e5a03e5\") " Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.293771 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrjwv\" (UniqueName: \"kubernetes.io/projected/b1d3e297-8696-43ae-84f8-da003e5a03e5-kube-api-access-mrjwv\") pod \"b1d3e297-8696-43ae-84f8-da003e5a03e5\" (UID: \"b1d3e297-8696-43ae-84f8-da003e5a03e5\") " Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.293880 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/096bd741-6394-4f58-9024-d425d96a0bd7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "096bd741-6394-4f58-9024-d425d96a0bd7" (UID: "096bd741-6394-4f58-9024-d425d96a0bd7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.294171 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1d3e297-8696-43ae-84f8-da003e5a03e5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b1d3e297-8696-43ae-84f8-da003e5a03e5" (UID: "b1d3e297-8696-43ae-84f8-da003e5a03e5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.294803 4559 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1d3e297-8696-43ae-84f8-da003e5a03e5-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.294828 4559 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79afaafd-5f51-4862-9c09-e7f817d54f73-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.294842 4559 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec53c68f-1637-47e8-8941-0ada991639ec-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.294852 4559 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.294866 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stqxc\" (UniqueName: \"kubernetes.io/projected/ec53c68f-1637-47e8-8941-0ada991639ec-kube-api-access-stqxc\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.294879 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztqns\" (UniqueName: \"kubernetes.io/projected/79afaafd-5f51-4862-9c09-e7f817d54f73-kube-api-access-ztqns\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.294890 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92rkt\" (UniqueName: \"kubernetes.io/projected/78fcad7c-4049-4dae-9fa1-2f8353447e12-kube-api-access-92rkt\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.294902 4559 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/096bd741-6394-4f58-9024-d425d96a0bd7-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.297414 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/096bd741-6394-4f58-9024-d425d96a0bd7-kube-api-access-b4zxb" (OuterVolumeSpecName: "kube-api-access-b4zxb") pod "096bd741-6394-4f58-9024-d425d96a0bd7" (UID: "096bd741-6394-4f58-9024-d425d96a0bd7"). InnerVolumeSpecName "kube-api-access-b4zxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.298242 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1d3e297-8696-43ae-84f8-da003e5a03e5-kube-api-access-mrjwv" (OuterVolumeSpecName: "kube-api-access-mrjwv") pod "b1d3e297-8696-43ae-84f8-da003e5a03e5" (UID: "b1d3e297-8696-43ae-84f8-da003e5a03e5"). InnerVolumeSpecName "kube-api-access-mrjwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.299796 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4-kube-api-access-jhclk" (OuterVolumeSpecName: "kube-api-access-jhclk") pod "fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4" (UID: "fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4"). InnerVolumeSpecName "kube-api-access-jhclk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.360980 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.397890 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrjwv\" (UniqueName: \"kubernetes.io/projected/b1d3e297-8696-43ae-84f8-da003e5a03e5-kube-api-access-mrjwv\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.397930 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4zxb\" (UniqueName: \"kubernetes.io/projected/096bd741-6394-4f58-9024-d425d96a0bd7-kube-api-access-b4zxb\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.397940 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhclk\" (UniqueName: \"kubernetes.io/projected/fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4-kube-api-access-jhclk\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.403140 4559 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="c3a4ae67-888b-46c2-bad1-92758aa2ea4f" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.143:9292/healthcheck\": read tcp 10.217.0.2:52026->10.217.0.143:9292: read: connection reset by peer" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.403448 4559 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="c3a4ae67-888b-46c2-bad1-92758aa2ea4f" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.143:9292/healthcheck\": read tcp 10.217.0.2:52028->10.217.0.143:9292: read: connection reset by peer" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.714212 4559 generic.go:334] "Generic (PLEG): container finished" podID="c3a4ae67-888b-46c2-bad1-92758aa2ea4f" containerID="27f8bbb0ad0a072b9c09d2f2502d82cb90133957a9ee3c475747e68d2e5c9159" exitCode=0 Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.714279 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c3a4ae67-888b-46c2-bad1-92758aa2ea4f","Type":"ContainerDied","Data":"27f8bbb0ad0a072b9c09d2f2502d82cb90133957a9ee3c475747e68d2e5c9159"} Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.718457 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-t898j" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.718454 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-t898j" event={"ID":"ec53c68f-1637-47e8-8941-0ada991639ec","Type":"ContainerDied","Data":"c4153b9d113e617e989c9c9cbbbcc00834cef21c275435c3ca5a77b53fb4700d"} Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.718519 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4153b9d113e617e989c9c9cbbbcc00834cef21c275435c3ca5a77b53fb4700d" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.720282 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-tp852" event={"ID":"b1d3e297-8696-43ae-84f8-da003e5a03e5","Type":"ContainerDied","Data":"a5b9fac63e0f0e877b763669fc0cec4cf4950d70a1fd7947641d7d13e9260d82"} Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.720332 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5b9fac63e0f0e877b763669fc0cec4cf4950d70a1fd7947641d7d13e9260d82" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.720421 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-tp852" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.724161 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-974e-account-create-qw2ds" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.724157 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-974e-account-create-qw2ds" event={"ID":"fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4","Type":"ContainerDied","Data":"8f031e2f3a5c14d534dd7b4ae0fb33c282d6f6ff97298f31bb734f9e3f7ac9a2"} Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.724265 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f031e2f3a5c14d534dd7b4ae0fb33c282d6f6ff97298f31bb734f9e3f7ac9a2" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.727449 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-d2d8-account-create-tfkjf" event={"ID":"78fcad7c-4049-4dae-9fa1-2f8353447e12","Type":"ContainerDied","Data":"d982566ecaf1d1196920ea57c167bb5c94062758e567d86d78eb32d05b68029f"} Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.727496 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d982566ecaf1d1196920ea57c167bb5c94062758e567d86d78eb32d05b68029f" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.727459 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-d2d8-account-create-tfkjf" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.728673 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-kktcm" event={"ID":"79afaafd-5f51-4862-9c09-e7f817d54f73","Type":"ContainerDied","Data":"b68662a211784edef9e35741c2a860eb3184e5a9ee5cce5b7e462ebc279c4622"} Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.728704 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b68662a211784edef9e35741c2a860eb3184e5a9ee5cce5b7e462ebc279c4622" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.728686 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kktcm" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.730706 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ade7-account-create-mj2zg" event={"ID":"096bd741-6394-4f58-9024-d425d96a0bd7","Type":"ContainerDied","Data":"c6c08850f922460ff6b39daa302fdb1bc6c430a0cf0324bb5b56e95a27442a08"} Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.730731 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6c08850f922460ff6b39daa302fdb1bc6c430a0cf0324bb5b56e95a27442a08" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.730763 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ade7-account-create-mj2zg" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.737006 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c3cf2c6-0694-425f-86e0-5313595a152a","Type":"ContainerStarted","Data":"c802573f831eb6d085fee6a6b4a69a700b50bdb41fc30adcf150eda529ffa8a4"} Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.738970 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa","Type":"ContainerStarted","Data":"90de052feb3f7b491c59e309bb5f565bb3e26d521862385e362c5d4c47f0e8f9"} Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.766915 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.145848574 podStartE2EDuration="14.766895689s" podCreationTimestamp="2025-11-23 07:00:24 +0000 UTC" firstStartedPulling="2025-11-23 07:00:25.322235196 +0000 UTC m=+927.344220811" lastFinishedPulling="2025-11-23 07:00:37.943282311 +0000 UTC m=+939.965267926" observedRunningTime="2025-11-23 07:00:38.757636485 +0000 UTC m=+940.779622099" watchObservedRunningTime="2025-11-23 07:00:38.766895689 +0000 UTC m=+940.788881303" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.787777 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.908696 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.908875 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-config-data\") pod \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.908988 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8n5d\" (UniqueName: \"kubernetes.io/projected/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-kube-api-access-b8n5d\") pod \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.909025 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-logs\") pod \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.909086 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-httpd-run\") pod \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.909210 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-public-tls-certs\") pod \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.909271 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-combined-ca-bundle\") pod \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.909421 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c3a4ae67-888b-46c2-bad1-92758aa2ea4f" (UID: "c3a4ae67-888b-46c2-bad1-92758aa2ea4f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.909514 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-logs" (OuterVolumeSpecName: "logs") pod "c3a4ae67-888b-46c2-bad1-92758aa2ea4f" (UID: "c3a4ae67-888b-46c2-bad1-92758aa2ea4f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.909695 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-scripts\") pod \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\" (UID: \"c3a4ae67-888b-46c2-bad1-92758aa2ea4f\") " Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.912208 4559 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.912259 4559 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.916942 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "c3a4ae67-888b-46c2-bad1-92758aa2ea4f" (UID: "c3a4ae67-888b-46c2-bad1-92758aa2ea4f"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.916981 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-kube-api-access-b8n5d" (OuterVolumeSpecName: "kube-api-access-b8n5d") pod "c3a4ae67-888b-46c2-bad1-92758aa2ea4f" (UID: "c3a4ae67-888b-46c2-bad1-92758aa2ea4f"). InnerVolumeSpecName "kube-api-access-b8n5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.917811 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-scripts" (OuterVolumeSpecName: "scripts") pod "c3a4ae67-888b-46c2-bad1-92758aa2ea4f" (UID: "c3a4ae67-888b-46c2-bad1-92758aa2ea4f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.937043 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c3a4ae67-888b-46c2-bad1-92758aa2ea4f" (UID: "c3a4ae67-888b-46c2-bad1-92758aa2ea4f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.950631 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c3a4ae67-888b-46c2-bad1-92758aa2ea4f" (UID: "c3a4ae67-888b-46c2-bad1-92758aa2ea4f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:38 crc kubenswrapper[4559]: I1123 07:00:38.953577 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-config-data" (OuterVolumeSpecName: "config-data") pod "c3a4ae67-888b-46c2-bad1-92758aa2ea4f" (UID: "c3a4ae67-888b-46c2-bad1-92758aa2ea4f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.014209 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.014239 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8n5d\" (UniqueName: \"kubernetes.io/projected/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-kube-api-access-b8n5d\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.014251 4559 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.014263 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.014272 4559 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3a4ae67-888b-46c2-bad1-92758aa2ea4f-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.014302 4559 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.031223 4559 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.116483 4559 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.597674 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.597880 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6b4fdd88f9-9hnng" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.748084 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.748057 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c3a4ae67-888b-46c2-bad1-92758aa2ea4f","Type":"ContainerDied","Data":"7c7c7c5fadb1108aa464a6cd385b659fc3953e6743d84fe1112bb7ec1fa288c9"} Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.748213 4559 scope.go:117] "RemoveContainer" containerID="27f8bbb0ad0a072b9c09d2f2502d82cb90133957a9ee3c475747e68d2e5c9159" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.754058 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c3cf2c6-0694-425f-86e0-5313595a152a","Type":"ContainerStarted","Data":"8e53f38f806286b0f4c7c750659d7bba842c1d122e03d888dd582faf67e2de3e"} Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.790424 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.803053 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.809577 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 07:00:39 crc kubenswrapper[4559]: E1123 07:00:39.810040 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4" containerName="mariadb-account-create" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.810061 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4" containerName="mariadb-account-create" Nov 23 07:00:39 crc kubenswrapper[4559]: E1123 07:00:39.810074 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="096bd741-6394-4f58-9024-d425d96a0bd7" containerName="mariadb-account-create" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.810082 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="096bd741-6394-4f58-9024-d425d96a0bd7" containerName="mariadb-account-create" Nov 23 07:00:39 crc kubenswrapper[4559]: E1123 07:00:39.810097 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec53c68f-1637-47e8-8941-0ada991639ec" containerName="mariadb-database-create" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.810104 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec53c68f-1637-47e8-8941-0ada991639ec" containerName="mariadb-database-create" Nov 23 07:00:39 crc kubenswrapper[4559]: E1123 07:00:39.810124 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79afaafd-5f51-4862-9c09-e7f817d54f73" containerName="mariadb-database-create" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.810131 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="79afaafd-5f51-4862-9c09-e7f817d54f73" containerName="mariadb-database-create" Nov 23 07:00:39 crc kubenswrapper[4559]: E1123 07:00:39.810419 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3a4ae67-888b-46c2-bad1-92758aa2ea4f" containerName="glance-log" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.810429 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3a4ae67-888b-46c2-bad1-92758aa2ea4f" containerName="glance-log" Nov 23 07:00:39 crc kubenswrapper[4559]: E1123 07:00:39.810469 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1d3e297-8696-43ae-84f8-da003e5a03e5" containerName="mariadb-database-create" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.810484 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1d3e297-8696-43ae-84f8-da003e5a03e5" containerName="mariadb-database-create" Nov 23 07:00:39 crc kubenswrapper[4559]: E1123 07:00:39.810496 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78fcad7c-4049-4dae-9fa1-2f8353447e12" containerName="mariadb-account-create" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.810503 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="78fcad7c-4049-4dae-9fa1-2f8353447e12" containerName="mariadb-account-create" Nov 23 07:00:39 crc kubenswrapper[4559]: E1123 07:00:39.810513 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3a4ae67-888b-46c2-bad1-92758aa2ea4f" containerName="glance-httpd" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.810521 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3a4ae67-888b-46c2-bad1-92758aa2ea4f" containerName="glance-httpd" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.810760 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="79afaafd-5f51-4862-9c09-e7f817d54f73" containerName="mariadb-database-create" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.810779 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3a4ae67-888b-46c2-bad1-92758aa2ea4f" containerName="glance-httpd" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.810793 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="78fcad7c-4049-4dae-9fa1-2f8353447e12" containerName="mariadb-account-create" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.810806 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec53c68f-1637-47e8-8941-0ada991639ec" containerName="mariadb-database-create" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.810818 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3a4ae67-888b-46c2-bad1-92758aa2ea4f" containerName="glance-log" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.810828 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1d3e297-8696-43ae-84f8-da003e5a03e5" containerName="mariadb-database-create" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.810841 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="096bd741-6394-4f58-9024-d425d96a0bd7" containerName="mariadb-account-create" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.810855 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4" containerName="mariadb-account-create" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.813427 4559 scope.go:117] "RemoveContainer" containerID="a87a2cb07397e7a692e575ea7ebc473172586952ba546154c48fc11d0195ff43" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.814524 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.816891 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.818160 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.823450 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.932920 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/24caf856-63a1-44eb-a0f2-7afc985ff668-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"24caf856-63a1-44eb-a0f2-7afc985ff668\") " pod="openstack/glance-default-external-api-0" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.933027 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"24caf856-63a1-44eb-a0f2-7afc985ff668\") " pod="openstack/glance-default-external-api-0" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.933232 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24caf856-63a1-44eb-a0f2-7afc985ff668-config-data\") pod \"glance-default-external-api-0\" (UID: \"24caf856-63a1-44eb-a0f2-7afc985ff668\") " pod="openstack/glance-default-external-api-0" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.933399 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24caf856-63a1-44eb-a0f2-7afc985ff668-logs\") pod \"glance-default-external-api-0\" (UID: \"24caf856-63a1-44eb-a0f2-7afc985ff668\") " pod="openstack/glance-default-external-api-0" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.934012 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24caf856-63a1-44eb-a0f2-7afc985ff668-scripts\") pod \"glance-default-external-api-0\" (UID: \"24caf856-63a1-44eb-a0f2-7afc985ff668\") " pod="openstack/glance-default-external-api-0" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.934819 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/24caf856-63a1-44eb-a0f2-7afc985ff668-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"24caf856-63a1-44eb-a0f2-7afc985ff668\") " pod="openstack/glance-default-external-api-0" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.934916 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24caf856-63a1-44eb-a0f2-7afc985ff668-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"24caf856-63a1-44eb-a0f2-7afc985ff668\") " pod="openstack/glance-default-external-api-0" Nov 23 07:00:39 crc kubenswrapper[4559]: I1123 07:00:39.934974 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9q7z\" (UniqueName: \"kubernetes.io/projected/24caf856-63a1-44eb-a0f2-7afc985ff668-kube-api-access-c9q7z\") pod \"glance-default-external-api-0\" (UID: \"24caf856-63a1-44eb-a0f2-7afc985ff668\") " pod="openstack/glance-default-external-api-0" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.037117 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/24caf856-63a1-44eb-a0f2-7afc985ff668-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"24caf856-63a1-44eb-a0f2-7afc985ff668\") " pod="openstack/glance-default-external-api-0" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.037213 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9q7z\" (UniqueName: \"kubernetes.io/projected/24caf856-63a1-44eb-a0f2-7afc985ff668-kube-api-access-c9q7z\") pod \"glance-default-external-api-0\" (UID: \"24caf856-63a1-44eb-a0f2-7afc985ff668\") " pod="openstack/glance-default-external-api-0" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.037240 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24caf856-63a1-44eb-a0f2-7afc985ff668-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"24caf856-63a1-44eb-a0f2-7afc985ff668\") " pod="openstack/glance-default-external-api-0" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.037292 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/24caf856-63a1-44eb-a0f2-7afc985ff668-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"24caf856-63a1-44eb-a0f2-7afc985ff668\") " pod="openstack/glance-default-external-api-0" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.037346 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"24caf856-63a1-44eb-a0f2-7afc985ff668\") " pod="openstack/glance-default-external-api-0" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.037478 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24caf856-63a1-44eb-a0f2-7afc985ff668-config-data\") pod \"glance-default-external-api-0\" (UID: \"24caf856-63a1-44eb-a0f2-7afc985ff668\") " pod="openstack/glance-default-external-api-0" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.037637 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24caf856-63a1-44eb-a0f2-7afc985ff668-logs\") pod \"glance-default-external-api-0\" (UID: \"24caf856-63a1-44eb-a0f2-7afc985ff668\") " pod="openstack/glance-default-external-api-0" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.037777 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24caf856-63a1-44eb-a0f2-7afc985ff668-scripts\") pod \"glance-default-external-api-0\" (UID: \"24caf856-63a1-44eb-a0f2-7afc985ff668\") " pod="openstack/glance-default-external-api-0" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.038240 4559 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"24caf856-63a1-44eb-a0f2-7afc985ff668\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-external-api-0" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.038608 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24caf856-63a1-44eb-a0f2-7afc985ff668-logs\") pod \"glance-default-external-api-0\" (UID: \"24caf856-63a1-44eb-a0f2-7afc985ff668\") " pod="openstack/glance-default-external-api-0" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.039297 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/24caf856-63a1-44eb-a0f2-7afc985ff668-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"24caf856-63a1-44eb-a0f2-7afc985ff668\") " pod="openstack/glance-default-external-api-0" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.043601 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24caf856-63a1-44eb-a0f2-7afc985ff668-config-data\") pod \"glance-default-external-api-0\" (UID: \"24caf856-63a1-44eb-a0f2-7afc985ff668\") " pod="openstack/glance-default-external-api-0" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.046055 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24caf856-63a1-44eb-a0f2-7afc985ff668-scripts\") pod \"glance-default-external-api-0\" (UID: \"24caf856-63a1-44eb-a0f2-7afc985ff668\") " pod="openstack/glance-default-external-api-0" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.052469 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24caf856-63a1-44eb-a0f2-7afc985ff668-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"24caf856-63a1-44eb-a0f2-7afc985ff668\") " pod="openstack/glance-default-external-api-0" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.053363 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/24caf856-63a1-44eb-a0f2-7afc985ff668-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"24caf856-63a1-44eb-a0f2-7afc985ff668\") " pod="openstack/glance-default-external-api-0" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.058244 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9q7z\" (UniqueName: \"kubernetes.io/projected/24caf856-63a1-44eb-a0f2-7afc985ff668-kube-api-access-c9q7z\") pod \"glance-default-external-api-0\" (UID: \"24caf856-63a1-44eb-a0f2-7afc985ff668\") " pod="openstack/glance-default-external-api-0" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.065350 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"24caf856-63a1-44eb-a0f2-7afc985ff668\") " pod="openstack/glance-default-external-api-0" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.179340 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.188498 4559 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="bf03ed6e-e623-4ee8-afd4-361d6d18097a" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.146:9292/healthcheck\": read tcp 10.217.0.2:51294->10.217.0.146:9292: read: connection reset by peer" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.188739 4559 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="bf03ed6e-e623-4ee8-afd4-361d6d18097a" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.146:9292/healthcheck\": read tcp 10.217.0.2:51284->10.217.0.146:9292: read: connection reset by peer" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.287107 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3a4ae67-888b-46c2-bad1-92758aa2ea4f" path="/var/lib/kubelet/pods/c3a4ae67-888b-46c2-bad1-92758aa2ea4f/volumes" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.683401 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 07:00:40 crc kubenswrapper[4559]: W1123 07:00:40.685276 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24caf856_63a1_44eb_a0f2_7afc985ff668.slice/crio-6622f0d027865b70433238b8627b6f4a54b635e654ee8ec52fea7809d7fb548c WatchSource:0}: Error finding container 6622f0d027865b70433238b8627b6f4a54b635e654ee8ec52fea7809d7fb548c: Status 404 returned error can't find the container with id 6622f0d027865b70433238b8627b6f4a54b635e654ee8ec52fea7809d7fb548c Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.709826 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.775114 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"24caf856-63a1-44eb-a0f2-7afc985ff668","Type":"ContainerStarted","Data":"6622f0d027865b70433238b8627b6f4a54b635e654ee8ec52fea7809d7fb548c"} Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.779872 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c3cf2c6-0694-425f-86e0-5313595a152a","Type":"ContainerStarted","Data":"718a61e88a6d582d90005b9e12c437b9d3794cc54688005f2183fe7f24a398fa"} Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.779924 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c3cf2c6-0694-425f-86e0-5313595a152a","Type":"ContainerStarted","Data":"f57e60d7bd4bf30ec8162ac39d46564fe3459bb4eaa13aa03c5906a600ea9429"} Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.784127 4559 generic.go:334] "Generic (PLEG): container finished" podID="bf03ed6e-e623-4ee8-afd4-361d6d18097a" containerID="cbd4d0e094e31047191103ed3e0e5d8a94e077ebe5cb31c5ec25c3b7850932a0" exitCode=0 Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.784171 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bf03ed6e-e623-4ee8-afd4-361d6d18097a","Type":"ContainerDied","Data":"cbd4d0e094e31047191103ed3e0e5d8a94e077ebe5cb31c5ec25c3b7850932a0"} Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.784194 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.784214 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bf03ed6e-e623-4ee8-afd4-361d6d18097a","Type":"ContainerDied","Data":"7f6109ea8478ed9d7e5550b148afb17ec76a652e3d7aa331882555826881c2c9"} Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.784234 4559 scope.go:117] "RemoveContainer" containerID="cbd4d0e094e31047191103ed3e0e5d8a94e077ebe5cb31c5ec25c3b7850932a0" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.815125 4559 scope.go:117] "RemoveContainer" containerID="a2503ed47292a8e094ec1cee0b8416e38dfefebe40c846cff1fd8c7958096c24" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.839963 4559 scope.go:117] "RemoveContainer" containerID="cbd4d0e094e31047191103ed3e0e5d8a94e077ebe5cb31c5ec25c3b7850932a0" Nov 23 07:00:40 crc kubenswrapper[4559]: E1123 07:00:40.845033 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbd4d0e094e31047191103ed3e0e5d8a94e077ebe5cb31c5ec25c3b7850932a0\": container with ID starting with cbd4d0e094e31047191103ed3e0e5d8a94e077ebe5cb31c5ec25c3b7850932a0 not found: ID does not exist" containerID="cbd4d0e094e31047191103ed3e0e5d8a94e077ebe5cb31c5ec25c3b7850932a0" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.845083 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbd4d0e094e31047191103ed3e0e5d8a94e077ebe5cb31c5ec25c3b7850932a0"} err="failed to get container status \"cbd4d0e094e31047191103ed3e0e5d8a94e077ebe5cb31c5ec25c3b7850932a0\": rpc error: code = NotFound desc = could not find container \"cbd4d0e094e31047191103ed3e0e5d8a94e077ebe5cb31c5ec25c3b7850932a0\": container with ID starting with cbd4d0e094e31047191103ed3e0e5d8a94e077ebe5cb31c5ec25c3b7850932a0 not found: ID does not exist" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.845121 4559 scope.go:117] "RemoveContainer" containerID="a2503ed47292a8e094ec1cee0b8416e38dfefebe40c846cff1fd8c7958096c24" Nov 23 07:00:40 crc kubenswrapper[4559]: E1123 07:00:40.845464 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2503ed47292a8e094ec1cee0b8416e38dfefebe40c846cff1fd8c7958096c24\": container with ID starting with a2503ed47292a8e094ec1cee0b8416e38dfefebe40c846cff1fd8c7958096c24 not found: ID does not exist" containerID="a2503ed47292a8e094ec1cee0b8416e38dfefebe40c846cff1fd8c7958096c24" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.845501 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2503ed47292a8e094ec1cee0b8416e38dfefebe40c846cff1fd8c7958096c24"} err="failed to get container status \"a2503ed47292a8e094ec1cee0b8416e38dfefebe40c846cff1fd8c7958096c24\": rpc error: code = NotFound desc = could not find container \"a2503ed47292a8e094ec1cee0b8416e38dfefebe40c846cff1fd8c7958096c24\": container with ID starting with a2503ed47292a8e094ec1cee0b8416e38dfefebe40c846cff1fd8c7958096c24 not found: ID does not exist" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.859545 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f455s\" (UniqueName: \"kubernetes.io/projected/bf03ed6e-e623-4ee8-afd4-361d6d18097a-kube-api-access-f455s\") pod \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.859621 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.859658 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf03ed6e-e623-4ee8-afd4-361d6d18097a-internal-tls-certs\") pod \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.860007 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf03ed6e-e623-4ee8-afd4-361d6d18097a-combined-ca-bundle\") pod \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.860161 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf03ed6e-e623-4ee8-afd4-361d6d18097a-scripts\") pod \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.860201 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf03ed6e-e623-4ee8-afd4-361d6d18097a-logs\") pod \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.860243 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bf03ed6e-e623-4ee8-afd4-361d6d18097a-httpd-run\") pod \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.860282 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf03ed6e-e623-4ee8-afd4-361d6d18097a-config-data\") pod \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\" (UID: \"bf03ed6e-e623-4ee8-afd4-361d6d18097a\") " Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.860779 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf03ed6e-e623-4ee8-afd4-361d6d18097a-logs" (OuterVolumeSpecName: "logs") pod "bf03ed6e-e623-4ee8-afd4-361d6d18097a" (UID: "bf03ed6e-e623-4ee8-afd4-361d6d18097a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.860815 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf03ed6e-e623-4ee8-afd4-361d6d18097a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "bf03ed6e-e623-4ee8-afd4-361d6d18097a" (UID: "bf03ed6e-e623-4ee8-afd4-361d6d18097a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.861274 4559 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf03ed6e-e623-4ee8-afd4-361d6d18097a-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.861294 4559 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bf03ed6e-e623-4ee8-afd4-361d6d18097a-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.864002 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf03ed6e-e623-4ee8-afd4-361d6d18097a-scripts" (OuterVolumeSpecName: "scripts") pod "bf03ed6e-e623-4ee8-afd4-361d6d18097a" (UID: "bf03ed6e-e623-4ee8-afd4-361d6d18097a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.864812 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf03ed6e-e623-4ee8-afd4-361d6d18097a-kube-api-access-f455s" (OuterVolumeSpecName: "kube-api-access-f455s") pod "bf03ed6e-e623-4ee8-afd4-361d6d18097a" (UID: "bf03ed6e-e623-4ee8-afd4-361d6d18097a"). InnerVolumeSpecName "kube-api-access-f455s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.866060 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "bf03ed6e-e623-4ee8-afd4-361d6d18097a" (UID: "bf03ed6e-e623-4ee8-afd4-361d6d18097a"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.891021 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf03ed6e-e623-4ee8-afd4-361d6d18097a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bf03ed6e-e623-4ee8-afd4-361d6d18097a" (UID: "bf03ed6e-e623-4ee8-afd4-361d6d18097a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.933156 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf03ed6e-e623-4ee8-afd4-361d6d18097a-config-data" (OuterVolumeSpecName: "config-data") pod "bf03ed6e-e623-4ee8-afd4-361d6d18097a" (UID: "bf03ed6e-e623-4ee8-afd4-361d6d18097a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.952827 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf03ed6e-e623-4ee8-afd4-361d6d18097a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "bf03ed6e-e623-4ee8-afd4-361d6d18097a" (UID: "bf03ed6e-e623-4ee8-afd4-361d6d18097a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.963242 4559 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf03ed6e-e623-4ee8-afd4-361d6d18097a-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.963272 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf03ed6e-e623-4ee8-afd4-361d6d18097a-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.963283 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f455s\" (UniqueName: \"kubernetes.io/projected/bf03ed6e-e623-4ee8-afd4-361d6d18097a-kube-api-access-f455s\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.963316 4559 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.963326 4559 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf03ed6e-e623-4ee8-afd4-361d6d18097a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.963336 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf03ed6e-e623-4ee8-afd4-361d6d18097a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:40 crc kubenswrapper[4559]: I1123 07:00:40.982176 4559 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.065118 4559 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.129356 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.135606 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.149673 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 07:00:41 crc kubenswrapper[4559]: E1123 07:00:41.150088 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf03ed6e-e623-4ee8-afd4-361d6d18097a" containerName="glance-log" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.150112 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf03ed6e-e623-4ee8-afd4-361d6d18097a" containerName="glance-log" Nov 23 07:00:41 crc kubenswrapper[4559]: E1123 07:00:41.150155 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf03ed6e-e623-4ee8-afd4-361d6d18097a" containerName="glance-httpd" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.150163 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf03ed6e-e623-4ee8-afd4-361d6d18097a" containerName="glance-httpd" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.150349 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf03ed6e-e623-4ee8-afd4-361d6d18097a" containerName="glance-httpd" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.150376 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf03ed6e-e623-4ee8-afd4-361d6d18097a" containerName="glance-log" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.151375 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.156461 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.156693 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.165165 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.269480 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c869f924-be1d-4e33-8a16-5a833a6ad9d0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c869f924-be1d-4e33-8a16-5a833a6ad9d0\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.270064 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c869f924-be1d-4e33-8a16-5a833a6ad9d0-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c869f924-be1d-4e33-8a16-5a833a6ad9d0\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.270183 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c869f924-be1d-4e33-8a16-5a833a6ad9d0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c869f924-be1d-4e33-8a16-5a833a6ad9d0\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.270299 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kfj8\" (UniqueName: \"kubernetes.io/projected/c869f924-be1d-4e33-8a16-5a833a6ad9d0-kube-api-access-5kfj8\") pod \"glance-default-internal-api-0\" (UID: \"c869f924-be1d-4e33-8a16-5a833a6ad9d0\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.270790 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c869f924-be1d-4e33-8a16-5a833a6ad9d0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c869f924-be1d-4e33-8a16-5a833a6ad9d0\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.270869 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"c869f924-be1d-4e33-8a16-5a833a6ad9d0\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.270931 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c869f924-be1d-4e33-8a16-5a833a6ad9d0-logs\") pod \"glance-default-internal-api-0\" (UID: \"c869f924-be1d-4e33-8a16-5a833a6ad9d0\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.271027 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c869f924-be1d-4e33-8a16-5a833a6ad9d0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c869f924-be1d-4e33-8a16-5a833a6ad9d0\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.374357 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c869f924-be1d-4e33-8a16-5a833a6ad9d0-logs\") pod \"glance-default-internal-api-0\" (UID: \"c869f924-be1d-4e33-8a16-5a833a6ad9d0\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.374867 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c869f924-be1d-4e33-8a16-5a833a6ad9d0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c869f924-be1d-4e33-8a16-5a833a6ad9d0\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.374905 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c869f924-be1d-4e33-8a16-5a833a6ad9d0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c869f924-be1d-4e33-8a16-5a833a6ad9d0\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.374966 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c869f924-be1d-4e33-8a16-5a833a6ad9d0-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c869f924-be1d-4e33-8a16-5a833a6ad9d0\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.374996 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c869f924-be1d-4e33-8a16-5a833a6ad9d0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c869f924-be1d-4e33-8a16-5a833a6ad9d0\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.375050 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kfj8\" (UniqueName: \"kubernetes.io/projected/c869f924-be1d-4e33-8a16-5a833a6ad9d0-kube-api-access-5kfj8\") pod \"glance-default-internal-api-0\" (UID: \"c869f924-be1d-4e33-8a16-5a833a6ad9d0\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.375094 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c869f924-be1d-4e33-8a16-5a833a6ad9d0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c869f924-be1d-4e33-8a16-5a833a6ad9d0\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.375137 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"c869f924-be1d-4e33-8a16-5a833a6ad9d0\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.377659 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c869f924-be1d-4e33-8a16-5a833a6ad9d0-logs\") pod \"glance-default-internal-api-0\" (UID: \"c869f924-be1d-4e33-8a16-5a833a6ad9d0\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.378883 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c869f924-be1d-4e33-8a16-5a833a6ad9d0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c869f924-be1d-4e33-8a16-5a833a6ad9d0\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.379108 4559 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"c869f924-be1d-4e33-8a16-5a833a6ad9d0\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.384029 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c869f924-be1d-4e33-8a16-5a833a6ad9d0-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c869f924-be1d-4e33-8a16-5a833a6ad9d0\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.384756 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c869f924-be1d-4e33-8a16-5a833a6ad9d0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c869f924-be1d-4e33-8a16-5a833a6ad9d0\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.384941 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c869f924-be1d-4e33-8a16-5a833a6ad9d0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c869f924-be1d-4e33-8a16-5a833a6ad9d0\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.390802 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c869f924-be1d-4e33-8a16-5a833a6ad9d0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c869f924-be1d-4e33-8a16-5a833a6ad9d0\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.400382 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kfj8\" (UniqueName: \"kubernetes.io/projected/c869f924-be1d-4e33-8a16-5a833a6ad9d0-kube-api-access-5kfj8\") pod \"glance-default-internal-api-0\" (UID: \"c869f924-be1d-4e33-8a16-5a833a6ad9d0\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.421297 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"c869f924-be1d-4e33-8a16-5a833a6ad9d0\") " pod="openstack/glance-default-internal-api-0" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.470731 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.778979 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7nhf7"] Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.781319 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7nhf7" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.783294 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.783709 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-fhsg9" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.783863 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.788082 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7nhf7"] Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.801635 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"24caf856-63a1-44eb-a0f2-7afc985ff668","Type":"ContainerStarted","Data":"129cae1201c6e0349ad0d90b1a918e920d63b3deb7e14f57197b73058d2a8470"} Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.801690 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"24caf856-63a1-44eb-a0f2-7afc985ff668","Type":"ContainerStarted","Data":"4caf7710e61a65d48533628fd03f1d6364682899536b4cf291c97fd0986d8669"} Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.833073 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=2.833054439 podStartE2EDuration="2.833054439s" podCreationTimestamp="2025-11-23 07:00:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:00:41.83085883 +0000 UTC m=+943.852844445" watchObservedRunningTime="2025-11-23 07:00:41.833054439 +0000 UTC m=+943.855040054" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.884657 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b08ced2f-75ec-4b10-9e87-e83572224054-scripts\") pod \"nova-cell0-conductor-db-sync-7nhf7\" (UID: \"b08ced2f-75ec-4b10-9e87-e83572224054\") " pod="openstack/nova-cell0-conductor-db-sync-7nhf7" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.884733 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b08ced2f-75ec-4b10-9e87-e83572224054-config-data\") pod \"nova-cell0-conductor-db-sync-7nhf7\" (UID: \"b08ced2f-75ec-4b10-9e87-e83572224054\") " pod="openstack/nova-cell0-conductor-db-sync-7nhf7" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.884929 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b08ced2f-75ec-4b10-9e87-e83572224054-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7nhf7\" (UID: \"b08ced2f-75ec-4b10-9e87-e83572224054\") " pod="openstack/nova-cell0-conductor-db-sync-7nhf7" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.886257 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fg557\" (UniqueName: \"kubernetes.io/projected/b08ced2f-75ec-4b10-9e87-e83572224054-kube-api-access-fg557\") pod \"nova-cell0-conductor-db-sync-7nhf7\" (UID: \"b08ced2f-75ec-4b10-9e87-e83572224054\") " pod="openstack/nova-cell0-conductor-db-sync-7nhf7" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.987529 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b08ced2f-75ec-4b10-9e87-e83572224054-scripts\") pod \"nova-cell0-conductor-db-sync-7nhf7\" (UID: \"b08ced2f-75ec-4b10-9e87-e83572224054\") " pod="openstack/nova-cell0-conductor-db-sync-7nhf7" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.987592 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b08ced2f-75ec-4b10-9e87-e83572224054-config-data\") pod \"nova-cell0-conductor-db-sync-7nhf7\" (UID: \"b08ced2f-75ec-4b10-9e87-e83572224054\") " pod="openstack/nova-cell0-conductor-db-sync-7nhf7" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.987701 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b08ced2f-75ec-4b10-9e87-e83572224054-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7nhf7\" (UID: \"b08ced2f-75ec-4b10-9e87-e83572224054\") " pod="openstack/nova-cell0-conductor-db-sync-7nhf7" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.987836 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fg557\" (UniqueName: \"kubernetes.io/projected/b08ced2f-75ec-4b10-9e87-e83572224054-kube-api-access-fg557\") pod \"nova-cell0-conductor-db-sync-7nhf7\" (UID: \"b08ced2f-75ec-4b10-9e87-e83572224054\") " pod="openstack/nova-cell0-conductor-db-sync-7nhf7" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.991020 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b08ced2f-75ec-4b10-9e87-e83572224054-scripts\") pod \"nova-cell0-conductor-db-sync-7nhf7\" (UID: \"b08ced2f-75ec-4b10-9e87-e83572224054\") " pod="openstack/nova-cell0-conductor-db-sync-7nhf7" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.991190 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b08ced2f-75ec-4b10-9e87-e83572224054-config-data\") pod \"nova-cell0-conductor-db-sync-7nhf7\" (UID: \"b08ced2f-75ec-4b10-9e87-e83572224054\") " pod="openstack/nova-cell0-conductor-db-sync-7nhf7" Nov 23 07:00:41 crc kubenswrapper[4559]: I1123 07:00:41.992238 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b08ced2f-75ec-4b10-9e87-e83572224054-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7nhf7\" (UID: \"b08ced2f-75ec-4b10-9e87-e83572224054\") " pod="openstack/nova-cell0-conductor-db-sync-7nhf7" Nov 23 07:00:42 crc kubenswrapper[4559]: I1123 07:00:42.008559 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fg557\" (UniqueName: \"kubernetes.io/projected/b08ced2f-75ec-4b10-9e87-e83572224054-kube-api-access-fg557\") pod \"nova-cell0-conductor-db-sync-7nhf7\" (UID: \"b08ced2f-75ec-4b10-9e87-e83572224054\") " pod="openstack/nova-cell0-conductor-db-sync-7nhf7" Nov 23 07:00:42 crc kubenswrapper[4559]: I1123 07:00:42.040262 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 07:00:42 crc kubenswrapper[4559]: I1123 07:00:42.114193 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7nhf7" Nov 23 07:00:42 crc kubenswrapper[4559]: I1123 07:00:42.288262 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf03ed6e-e623-4ee8-afd4-361d6d18097a" path="/var/lib/kubelet/pods/bf03ed6e-e623-4ee8-afd4-361d6d18097a/volumes" Nov 23 07:00:42 crc kubenswrapper[4559]: I1123 07:00:42.620685 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7nhf7"] Nov 23 07:00:42 crc kubenswrapper[4559]: I1123 07:00:42.824002 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c3cf2c6-0694-425f-86e0-5313595a152a","Type":"ContainerStarted","Data":"223f76e6644867f7ccc114128f10cd1afe42c1dd8311f12a81b59d7b08961633"} Nov 23 07:00:42 crc kubenswrapper[4559]: I1123 07:00:42.824171 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 23 07:00:42 crc kubenswrapper[4559]: I1123 07:00:42.824231 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9c3cf2c6-0694-425f-86e0-5313595a152a" containerName="ceilometer-central-agent" containerID="cri-o://8e53f38f806286b0f4c7c750659d7bba842c1d122e03d888dd582faf67e2de3e" gracePeriod=30 Nov 23 07:00:42 crc kubenswrapper[4559]: I1123 07:00:42.824283 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9c3cf2c6-0694-425f-86e0-5313595a152a" containerName="ceilometer-notification-agent" containerID="cri-o://f57e60d7bd4bf30ec8162ac39d46564fe3459bb4eaa13aa03c5906a600ea9429" gracePeriod=30 Nov 23 07:00:42 crc kubenswrapper[4559]: I1123 07:00:42.824253 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9c3cf2c6-0694-425f-86e0-5313595a152a" containerName="sg-core" containerID="cri-o://718a61e88a6d582d90005b9e12c437b9d3794cc54688005f2183fe7f24a398fa" gracePeriod=30 Nov 23 07:00:42 crc kubenswrapper[4559]: I1123 07:00:42.824266 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9c3cf2c6-0694-425f-86e0-5313595a152a" containerName="proxy-httpd" containerID="cri-o://223f76e6644867f7ccc114128f10cd1afe42c1dd8311f12a81b59d7b08961633" gracePeriod=30 Nov 23 07:00:42 crc kubenswrapper[4559]: I1123 07:00:42.835381 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7nhf7" event={"ID":"b08ced2f-75ec-4b10-9e87-e83572224054","Type":"ContainerStarted","Data":"01def69877136d8d0f89c9f6864060b5f83389e44927a26f3a24e634aecf55f6"} Nov 23 07:00:42 crc kubenswrapper[4559]: I1123 07:00:42.837313 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c869f924-be1d-4e33-8a16-5a833a6ad9d0","Type":"ContainerStarted","Data":"c405f3859683170bf135915b6edd18481c98c86e851c971bdf650fc96e2b689c"} Nov 23 07:00:42 crc kubenswrapper[4559]: I1123 07:00:42.837340 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c869f924-be1d-4e33-8a16-5a833a6ad9d0","Type":"ContainerStarted","Data":"5f10b32116d5a72e90dd86052b852d7534cffe8501e538f998f2b6589a1e180a"} Nov 23 07:00:42 crc kubenswrapper[4559]: I1123 07:00:42.852777 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=6.2590150399999995 podStartE2EDuration="9.852763899s" podCreationTimestamp="2025-11-23 07:00:33 +0000 UTC" firstStartedPulling="2025-11-23 07:00:38.357465234 +0000 UTC m=+940.379450849" lastFinishedPulling="2025-11-23 07:00:41.951214095 +0000 UTC m=+943.973199708" observedRunningTime="2025-11-23 07:00:42.844450914 +0000 UTC m=+944.866436528" watchObservedRunningTime="2025-11-23 07:00:42.852763899 +0000 UTC m=+944.874749512" Nov 23 07:00:43 crc kubenswrapper[4559]: I1123 07:00:43.848938 4559 generic.go:334] "Generic (PLEG): container finished" podID="9c3cf2c6-0694-425f-86e0-5313595a152a" containerID="223f76e6644867f7ccc114128f10cd1afe42c1dd8311f12a81b59d7b08961633" exitCode=0 Nov 23 07:00:43 crc kubenswrapper[4559]: I1123 07:00:43.849310 4559 generic.go:334] "Generic (PLEG): container finished" podID="9c3cf2c6-0694-425f-86e0-5313595a152a" containerID="718a61e88a6d582d90005b9e12c437b9d3794cc54688005f2183fe7f24a398fa" exitCode=2 Nov 23 07:00:43 crc kubenswrapper[4559]: I1123 07:00:43.849018 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c3cf2c6-0694-425f-86e0-5313595a152a","Type":"ContainerDied","Data":"223f76e6644867f7ccc114128f10cd1afe42c1dd8311f12a81b59d7b08961633"} Nov 23 07:00:43 crc kubenswrapper[4559]: I1123 07:00:43.849321 4559 generic.go:334] "Generic (PLEG): container finished" podID="9c3cf2c6-0694-425f-86e0-5313595a152a" containerID="f57e60d7bd4bf30ec8162ac39d46564fe3459bb4eaa13aa03c5906a600ea9429" exitCode=0 Nov 23 07:00:43 crc kubenswrapper[4559]: I1123 07:00:43.849379 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c3cf2c6-0694-425f-86e0-5313595a152a","Type":"ContainerDied","Data":"718a61e88a6d582d90005b9e12c437b9d3794cc54688005f2183fe7f24a398fa"} Nov 23 07:00:43 crc kubenswrapper[4559]: I1123 07:00:43.849398 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c3cf2c6-0694-425f-86e0-5313595a152a","Type":"ContainerDied","Data":"f57e60d7bd4bf30ec8162ac39d46564fe3459bb4eaa13aa03c5906a600ea9429"} Nov 23 07:00:43 crc kubenswrapper[4559]: I1123 07:00:43.851610 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c869f924-be1d-4e33-8a16-5a833a6ad9d0","Type":"ContainerStarted","Data":"f31d65e0db921121cc0713be53d39e61e072b7e4708058eb060c71b025f2b3c2"} Nov 23 07:00:43 crc kubenswrapper[4559]: I1123 07:00:43.878240 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.878224345 podStartE2EDuration="2.878224345s" podCreationTimestamp="2025-11-23 07:00:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:00:43.871409818 +0000 UTC m=+945.893395432" watchObservedRunningTime="2025-11-23 07:00:43.878224345 +0000 UTC m=+945.900209958" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.593403 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.698098 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9c3cf2c6-0694-425f-86e0-5313595a152a-sg-core-conf-yaml\") pod \"9c3cf2c6-0694-425f-86e0-5313595a152a\" (UID: \"9c3cf2c6-0694-425f-86e0-5313595a152a\") " Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.698133 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c3cf2c6-0694-425f-86e0-5313595a152a-config-data\") pod \"9c3cf2c6-0694-425f-86e0-5313595a152a\" (UID: \"9c3cf2c6-0694-425f-86e0-5313595a152a\") " Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.698177 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c3cf2c6-0694-425f-86e0-5313595a152a-combined-ca-bundle\") pod \"9c3cf2c6-0694-425f-86e0-5313595a152a\" (UID: \"9c3cf2c6-0694-425f-86e0-5313595a152a\") " Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.698231 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qws98\" (UniqueName: \"kubernetes.io/projected/9c3cf2c6-0694-425f-86e0-5313595a152a-kube-api-access-qws98\") pod \"9c3cf2c6-0694-425f-86e0-5313595a152a\" (UID: \"9c3cf2c6-0694-425f-86e0-5313595a152a\") " Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.698290 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c3cf2c6-0694-425f-86e0-5313595a152a-run-httpd\") pod \"9c3cf2c6-0694-425f-86e0-5313595a152a\" (UID: \"9c3cf2c6-0694-425f-86e0-5313595a152a\") " Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.698354 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c3cf2c6-0694-425f-86e0-5313595a152a-log-httpd\") pod \"9c3cf2c6-0694-425f-86e0-5313595a152a\" (UID: \"9c3cf2c6-0694-425f-86e0-5313595a152a\") " Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.698370 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c3cf2c6-0694-425f-86e0-5313595a152a-scripts\") pod \"9c3cf2c6-0694-425f-86e0-5313595a152a\" (UID: \"9c3cf2c6-0694-425f-86e0-5313595a152a\") " Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.699367 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c3cf2c6-0694-425f-86e0-5313595a152a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9c3cf2c6-0694-425f-86e0-5313595a152a" (UID: "9c3cf2c6-0694-425f-86e0-5313595a152a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.699437 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c3cf2c6-0694-425f-86e0-5313595a152a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9c3cf2c6-0694-425f-86e0-5313595a152a" (UID: "9c3cf2c6-0694-425f-86e0-5313595a152a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.705290 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c3cf2c6-0694-425f-86e0-5313595a152a-kube-api-access-qws98" (OuterVolumeSpecName: "kube-api-access-qws98") pod "9c3cf2c6-0694-425f-86e0-5313595a152a" (UID: "9c3cf2c6-0694-425f-86e0-5313595a152a"). InnerVolumeSpecName "kube-api-access-qws98". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.714742 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c3cf2c6-0694-425f-86e0-5313595a152a-scripts" (OuterVolumeSpecName: "scripts") pod "9c3cf2c6-0694-425f-86e0-5313595a152a" (UID: "9c3cf2c6-0694-425f-86e0-5313595a152a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.722941 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c3cf2c6-0694-425f-86e0-5313595a152a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9c3cf2c6-0694-425f-86e0-5313595a152a" (UID: "9c3cf2c6-0694-425f-86e0-5313595a152a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.757081 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c3cf2c6-0694-425f-86e0-5313595a152a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c3cf2c6-0694-425f-86e0-5313595a152a" (UID: "9c3cf2c6-0694-425f-86e0-5313595a152a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.772804 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c3cf2c6-0694-425f-86e0-5313595a152a-config-data" (OuterVolumeSpecName: "config-data") pod "9c3cf2c6-0694-425f-86e0-5313595a152a" (UID: "9c3cf2c6-0694-425f-86e0-5313595a152a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.801137 4559 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c3cf2c6-0694-425f-86e0-5313595a152a-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.801163 4559 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c3cf2c6-0694-425f-86e0-5313595a152a-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.801174 4559 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c3cf2c6-0694-425f-86e0-5313595a152a-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.801181 4559 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9c3cf2c6-0694-425f-86e0-5313595a152a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.801191 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c3cf2c6-0694-425f-86e0-5313595a152a-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.801199 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c3cf2c6-0694-425f-86e0-5313595a152a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.801207 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qws98\" (UniqueName: \"kubernetes.io/projected/9c3cf2c6-0694-425f-86e0-5313595a152a-kube-api-access-qws98\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.893511 4559 generic.go:334] "Generic (PLEG): container finished" podID="9c3cf2c6-0694-425f-86e0-5313595a152a" containerID="8e53f38f806286b0f4c7c750659d7bba842c1d122e03d888dd582faf67e2de3e" exitCode=0 Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.893614 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c3cf2c6-0694-425f-86e0-5313595a152a","Type":"ContainerDied","Data":"8e53f38f806286b0f4c7c750659d7bba842c1d122e03d888dd582faf67e2de3e"} Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.893638 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.893856 4559 scope.go:117] "RemoveContainer" containerID="223f76e6644867f7ccc114128f10cd1afe42c1dd8311f12a81b59d7b08961633" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.893780 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c3cf2c6-0694-425f-86e0-5313595a152a","Type":"ContainerDied","Data":"c802573f831eb6d085fee6a6b4a69a700b50bdb41fc30adcf150eda529ffa8a4"} Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.925507 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.932892 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.952319 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:00:47 crc kubenswrapper[4559]: E1123 07:00:47.952809 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c3cf2c6-0694-425f-86e0-5313595a152a" containerName="ceilometer-notification-agent" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.952827 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c3cf2c6-0694-425f-86e0-5313595a152a" containerName="ceilometer-notification-agent" Nov 23 07:00:47 crc kubenswrapper[4559]: E1123 07:00:47.952840 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c3cf2c6-0694-425f-86e0-5313595a152a" containerName="ceilometer-central-agent" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.952846 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c3cf2c6-0694-425f-86e0-5313595a152a" containerName="ceilometer-central-agent" Nov 23 07:00:47 crc kubenswrapper[4559]: E1123 07:00:47.952864 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c3cf2c6-0694-425f-86e0-5313595a152a" containerName="sg-core" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.952871 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c3cf2c6-0694-425f-86e0-5313595a152a" containerName="sg-core" Nov 23 07:00:47 crc kubenswrapper[4559]: E1123 07:00:47.952879 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c3cf2c6-0694-425f-86e0-5313595a152a" containerName="proxy-httpd" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.952884 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c3cf2c6-0694-425f-86e0-5313595a152a" containerName="proxy-httpd" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.953137 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c3cf2c6-0694-425f-86e0-5313595a152a" containerName="sg-core" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.953152 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c3cf2c6-0694-425f-86e0-5313595a152a" containerName="proxy-httpd" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.953161 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c3cf2c6-0694-425f-86e0-5313595a152a" containerName="ceilometer-central-agent" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.953182 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c3cf2c6-0694-425f-86e0-5313595a152a" containerName="ceilometer-notification-agent" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.954941 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.959037 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.959408 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 23 07:00:47 crc kubenswrapper[4559]: I1123 07:00:47.968170 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:00:48 crc kubenswrapper[4559]: I1123 07:00:48.006180 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-config-data\") pod \"ceilometer-0\" (UID: \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\") " pod="openstack/ceilometer-0" Nov 23 07:00:48 crc kubenswrapper[4559]: I1123 07:00:48.006443 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\") " pod="openstack/ceilometer-0" Nov 23 07:00:48 crc kubenswrapper[4559]: I1123 07:00:48.006507 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-run-httpd\") pod \"ceilometer-0\" (UID: \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\") " pod="openstack/ceilometer-0" Nov 23 07:00:48 crc kubenswrapper[4559]: I1123 07:00:48.006622 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-log-httpd\") pod \"ceilometer-0\" (UID: \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\") " pod="openstack/ceilometer-0" Nov 23 07:00:48 crc kubenswrapper[4559]: I1123 07:00:48.006721 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cn6cg\" (UniqueName: \"kubernetes.io/projected/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-kube-api-access-cn6cg\") pod \"ceilometer-0\" (UID: \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\") " pod="openstack/ceilometer-0" Nov 23 07:00:48 crc kubenswrapper[4559]: I1123 07:00:48.006857 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-scripts\") pod \"ceilometer-0\" (UID: \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\") " pod="openstack/ceilometer-0" Nov 23 07:00:48 crc kubenswrapper[4559]: I1123 07:00:48.006879 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\") " pod="openstack/ceilometer-0" Nov 23 07:00:48 crc kubenswrapper[4559]: I1123 07:00:48.109607 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\") " pod="openstack/ceilometer-0" Nov 23 07:00:48 crc kubenswrapper[4559]: I1123 07:00:48.109666 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-run-httpd\") pod \"ceilometer-0\" (UID: \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\") " pod="openstack/ceilometer-0" Nov 23 07:00:48 crc kubenswrapper[4559]: I1123 07:00:48.109716 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-log-httpd\") pod \"ceilometer-0\" (UID: \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\") " pod="openstack/ceilometer-0" Nov 23 07:00:48 crc kubenswrapper[4559]: I1123 07:00:48.109750 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cn6cg\" (UniqueName: \"kubernetes.io/projected/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-kube-api-access-cn6cg\") pod \"ceilometer-0\" (UID: \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\") " pod="openstack/ceilometer-0" Nov 23 07:00:48 crc kubenswrapper[4559]: I1123 07:00:48.109803 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-scripts\") pod \"ceilometer-0\" (UID: \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\") " pod="openstack/ceilometer-0" Nov 23 07:00:48 crc kubenswrapper[4559]: I1123 07:00:48.109825 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\") " pod="openstack/ceilometer-0" Nov 23 07:00:48 crc kubenswrapper[4559]: I1123 07:00:48.109889 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-config-data\") pod \"ceilometer-0\" (UID: \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\") " pod="openstack/ceilometer-0" Nov 23 07:00:48 crc kubenswrapper[4559]: I1123 07:00:48.110573 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-run-httpd\") pod \"ceilometer-0\" (UID: \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\") " pod="openstack/ceilometer-0" Nov 23 07:00:48 crc kubenswrapper[4559]: I1123 07:00:48.110730 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-log-httpd\") pod \"ceilometer-0\" (UID: \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\") " pod="openstack/ceilometer-0" Nov 23 07:00:48 crc kubenswrapper[4559]: I1123 07:00:48.114964 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-config-data\") pod \"ceilometer-0\" (UID: \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\") " pod="openstack/ceilometer-0" Nov 23 07:00:48 crc kubenswrapper[4559]: I1123 07:00:48.115385 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-scripts\") pod \"ceilometer-0\" (UID: \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\") " pod="openstack/ceilometer-0" Nov 23 07:00:48 crc kubenswrapper[4559]: I1123 07:00:48.116020 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\") " pod="openstack/ceilometer-0" Nov 23 07:00:48 crc kubenswrapper[4559]: I1123 07:00:48.118065 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\") " pod="openstack/ceilometer-0" Nov 23 07:00:48 crc kubenswrapper[4559]: I1123 07:00:48.125725 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cn6cg\" (UniqueName: \"kubernetes.io/projected/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-kube-api-access-cn6cg\") pod \"ceilometer-0\" (UID: \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\") " pod="openstack/ceilometer-0" Nov 23 07:00:48 crc kubenswrapper[4559]: I1123 07:00:48.276827 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:00:48 crc kubenswrapper[4559]: I1123 07:00:48.299416 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c3cf2c6-0694-425f-86e0-5313595a152a" path="/var/lib/kubelet/pods/9c3cf2c6-0694-425f-86e0-5313595a152a/volumes" Nov 23 07:00:50 crc kubenswrapper[4559]: I1123 07:00:50.180108 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 23 07:00:50 crc kubenswrapper[4559]: I1123 07:00:50.180541 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 23 07:00:50 crc kubenswrapper[4559]: I1123 07:00:50.216249 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 23 07:00:50 crc kubenswrapper[4559]: I1123 07:00:50.233612 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 23 07:00:50 crc kubenswrapper[4559]: I1123 07:00:50.928497 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 23 07:00:50 crc kubenswrapper[4559]: I1123 07:00:50.928548 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 23 07:00:51 crc kubenswrapper[4559]: I1123 07:00:51.069975 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:00:51 crc kubenswrapper[4559]: I1123 07:00:51.240423 4559 scope.go:117] "RemoveContainer" containerID="718a61e88a6d582d90005b9e12c437b9d3794cc54688005f2183fe7f24a398fa" Nov 23 07:00:51 crc kubenswrapper[4559]: I1123 07:00:51.274525 4559 scope.go:117] "RemoveContainer" containerID="f57e60d7bd4bf30ec8162ac39d46564fe3459bb4eaa13aa03c5906a600ea9429" Nov 23 07:00:51 crc kubenswrapper[4559]: I1123 07:00:51.294681 4559 scope.go:117] "RemoveContainer" containerID="8e53f38f806286b0f4c7c750659d7bba842c1d122e03d888dd582faf67e2de3e" Nov 23 07:00:51 crc kubenswrapper[4559]: I1123 07:00:51.434771 4559 scope.go:117] "RemoveContainer" containerID="223f76e6644867f7ccc114128f10cd1afe42c1dd8311f12a81b59d7b08961633" Nov 23 07:00:51 crc kubenswrapper[4559]: E1123 07:00:51.435853 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"223f76e6644867f7ccc114128f10cd1afe42c1dd8311f12a81b59d7b08961633\": container with ID starting with 223f76e6644867f7ccc114128f10cd1afe42c1dd8311f12a81b59d7b08961633 not found: ID does not exist" containerID="223f76e6644867f7ccc114128f10cd1afe42c1dd8311f12a81b59d7b08961633" Nov 23 07:00:51 crc kubenswrapper[4559]: I1123 07:00:51.435900 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"223f76e6644867f7ccc114128f10cd1afe42c1dd8311f12a81b59d7b08961633"} err="failed to get container status \"223f76e6644867f7ccc114128f10cd1afe42c1dd8311f12a81b59d7b08961633\": rpc error: code = NotFound desc = could not find container \"223f76e6644867f7ccc114128f10cd1afe42c1dd8311f12a81b59d7b08961633\": container with ID starting with 223f76e6644867f7ccc114128f10cd1afe42c1dd8311f12a81b59d7b08961633 not found: ID does not exist" Nov 23 07:00:51 crc kubenswrapper[4559]: I1123 07:00:51.435926 4559 scope.go:117] "RemoveContainer" containerID="718a61e88a6d582d90005b9e12c437b9d3794cc54688005f2183fe7f24a398fa" Nov 23 07:00:51 crc kubenswrapper[4559]: E1123 07:00:51.436323 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"718a61e88a6d582d90005b9e12c437b9d3794cc54688005f2183fe7f24a398fa\": container with ID starting with 718a61e88a6d582d90005b9e12c437b9d3794cc54688005f2183fe7f24a398fa not found: ID does not exist" containerID="718a61e88a6d582d90005b9e12c437b9d3794cc54688005f2183fe7f24a398fa" Nov 23 07:00:51 crc kubenswrapper[4559]: I1123 07:00:51.436342 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"718a61e88a6d582d90005b9e12c437b9d3794cc54688005f2183fe7f24a398fa"} err="failed to get container status \"718a61e88a6d582d90005b9e12c437b9d3794cc54688005f2183fe7f24a398fa\": rpc error: code = NotFound desc = could not find container \"718a61e88a6d582d90005b9e12c437b9d3794cc54688005f2183fe7f24a398fa\": container with ID starting with 718a61e88a6d582d90005b9e12c437b9d3794cc54688005f2183fe7f24a398fa not found: ID does not exist" Nov 23 07:00:51 crc kubenswrapper[4559]: I1123 07:00:51.436358 4559 scope.go:117] "RemoveContainer" containerID="f57e60d7bd4bf30ec8162ac39d46564fe3459bb4eaa13aa03c5906a600ea9429" Nov 23 07:00:51 crc kubenswrapper[4559]: E1123 07:00:51.436536 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f57e60d7bd4bf30ec8162ac39d46564fe3459bb4eaa13aa03c5906a600ea9429\": container with ID starting with f57e60d7bd4bf30ec8162ac39d46564fe3459bb4eaa13aa03c5906a600ea9429 not found: ID does not exist" containerID="f57e60d7bd4bf30ec8162ac39d46564fe3459bb4eaa13aa03c5906a600ea9429" Nov 23 07:00:51 crc kubenswrapper[4559]: I1123 07:00:51.436556 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f57e60d7bd4bf30ec8162ac39d46564fe3459bb4eaa13aa03c5906a600ea9429"} err="failed to get container status \"f57e60d7bd4bf30ec8162ac39d46564fe3459bb4eaa13aa03c5906a600ea9429\": rpc error: code = NotFound desc = could not find container \"f57e60d7bd4bf30ec8162ac39d46564fe3459bb4eaa13aa03c5906a600ea9429\": container with ID starting with f57e60d7bd4bf30ec8162ac39d46564fe3459bb4eaa13aa03c5906a600ea9429 not found: ID does not exist" Nov 23 07:00:51 crc kubenswrapper[4559]: I1123 07:00:51.436567 4559 scope.go:117] "RemoveContainer" containerID="8e53f38f806286b0f4c7c750659d7bba842c1d122e03d888dd582faf67e2de3e" Nov 23 07:00:51 crc kubenswrapper[4559]: E1123 07:00:51.437696 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e53f38f806286b0f4c7c750659d7bba842c1d122e03d888dd582faf67e2de3e\": container with ID starting with 8e53f38f806286b0f4c7c750659d7bba842c1d122e03d888dd582faf67e2de3e not found: ID does not exist" containerID="8e53f38f806286b0f4c7c750659d7bba842c1d122e03d888dd582faf67e2de3e" Nov 23 07:00:51 crc kubenswrapper[4559]: I1123 07:00:51.437721 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e53f38f806286b0f4c7c750659d7bba842c1d122e03d888dd582faf67e2de3e"} err="failed to get container status \"8e53f38f806286b0f4c7c750659d7bba842c1d122e03d888dd582faf67e2de3e\": rpc error: code = NotFound desc = could not find container \"8e53f38f806286b0f4c7c750659d7bba842c1d122e03d888dd582faf67e2de3e\": container with ID starting with 8e53f38f806286b0f4c7c750659d7bba842c1d122e03d888dd582faf67e2de3e not found: ID does not exist" Nov 23 07:00:51 crc kubenswrapper[4559]: I1123 07:00:51.471765 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 23 07:00:51 crc kubenswrapper[4559]: I1123 07:00:51.471837 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 23 07:00:51 crc kubenswrapper[4559]: I1123 07:00:51.505436 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 23 07:00:51 crc kubenswrapper[4559]: I1123 07:00:51.517968 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 23 07:00:51 crc kubenswrapper[4559]: I1123 07:00:51.733825 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:00:51 crc kubenswrapper[4559]: I1123 07:00:51.938898 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7nhf7" event={"ID":"b08ced2f-75ec-4b10-9e87-e83572224054","Type":"ContainerStarted","Data":"dd9cc55d2d749115f38ba72ac2c037e554c026fa3b9dfb051ed05730645e4c3b"} Nov 23 07:00:51 crc kubenswrapper[4559]: I1123 07:00:51.940591 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5","Type":"ContainerStarted","Data":"213f547a280bfd1fcd1151aee3b760e8c7deaafc6f594ce952ad900811276990"} Nov 23 07:00:51 crc kubenswrapper[4559]: I1123 07:00:51.943696 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 23 07:00:51 crc kubenswrapper[4559]: I1123 07:00:51.944001 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 23 07:00:51 crc kubenswrapper[4559]: I1123 07:00:51.963416 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-7nhf7" podStartSLOduration=2.291845389 podStartE2EDuration="10.963394778s" podCreationTimestamp="2025-11-23 07:00:41 +0000 UTC" firstStartedPulling="2025-11-23 07:00:42.634178501 +0000 UTC m=+944.656164115" lastFinishedPulling="2025-11-23 07:00:51.305727891 +0000 UTC m=+953.327713504" observedRunningTime="2025-11-23 07:00:51.955614717 +0000 UTC m=+953.977600331" watchObservedRunningTime="2025-11-23 07:00:51.963394778 +0000 UTC m=+953.985380392" Nov 23 07:00:52 crc kubenswrapper[4559]: I1123 07:00:52.685745 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 23 07:00:52 crc kubenswrapper[4559]: I1123 07:00:52.714530 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 23 07:00:52 crc kubenswrapper[4559]: I1123 07:00:52.953153 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5","Type":"ContainerStarted","Data":"3430ff2eaf1f572f57a8701d797eb8c96b432ead2047c1b48c5fc67ecbc32953"} Nov 23 07:00:53 crc kubenswrapper[4559]: I1123 07:00:53.745108 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 23 07:00:53 crc kubenswrapper[4559]: I1123 07:00:53.960781 4559 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 23 07:00:53 crc kubenswrapper[4559]: I1123 07:00:53.961726 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5","Type":"ContainerStarted","Data":"e306043b6c63645e4301c6d2ada6b062eb31ed1497ee6886297d2919099fddce"} Nov 23 07:00:54 crc kubenswrapper[4559]: I1123 07:00:54.081164 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 23 07:00:54 crc kubenswrapper[4559]: I1123 07:00:54.974731 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5","Type":"ContainerStarted","Data":"8a0cc150042c2d08c3205b6303843fb4a3135a34fbf86beff1d50658f99f4754"} Nov 23 07:00:55 crc kubenswrapper[4559]: I1123 07:00:55.987326 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5","Type":"ContainerStarted","Data":"8bd770cea540459a48d5d916eefc70b323ab4befddb41271517ac7d3c56fd34e"} Nov 23 07:00:55 crc kubenswrapper[4559]: I1123 07:00:55.987485 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="665d6fe5-5f01-48f1-bbe3-1ee83666e9d5" containerName="ceilometer-central-agent" containerID="cri-o://3430ff2eaf1f572f57a8701d797eb8c96b432ead2047c1b48c5fc67ecbc32953" gracePeriod=30 Nov 23 07:00:55 crc kubenswrapper[4559]: I1123 07:00:55.987704 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="665d6fe5-5f01-48f1-bbe3-1ee83666e9d5" containerName="sg-core" containerID="cri-o://8a0cc150042c2d08c3205b6303843fb4a3135a34fbf86beff1d50658f99f4754" gracePeriod=30 Nov 23 07:00:55 crc kubenswrapper[4559]: I1123 07:00:55.987675 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="665d6fe5-5f01-48f1-bbe3-1ee83666e9d5" containerName="ceilometer-notification-agent" containerID="cri-o://e306043b6c63645e4301c6d2ada6b062eb31ed1497ee6886297d2919099fddce" gracePeriod=30 Nov 23 07:00:55 crc kubenswrapper[4559]: I1123 07:00:55.987553 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="665d6fe5-5f01-48f1-bbe3-1ee83666e9d5" containerName="proxy-httpd" containerID="cri-o://8bd770cea540459a48d5d916eefc70b323ab4befddb41271517ac7d3c56fd34e" gracePeriod=30 Nov 23 07:00:55 crc kubenswrapper[4559]: I1123 07:00:55.988537 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 23 07:00:56 crc kubenswrapper[4559]: I1123 07:00:56.998579 4559 generic.go:334] "Generic (PLEG): container finished" podID="665d6fe5-5f01-48f1-bbe3-1ee83666e9d5" containerID="8bd770cea540459a48d5d916eefc70b323ab4befddb41271517ac7d3c56fd34e" exitCode=0 Nov 23 07:00:56 crc kubenswrapper[4559]: I1123 07:00:56.998936 4559 generic.go:334] "Generic (PLEG): container finished" podID="665d6fe5-5f01-48f1-bbe3-1ee83666e9d5" containerID="8a0cc150042c2d08c3205b6303843fb4a3135a34fbf86beff1d50658f99f4754" exitCode=2 Nov 23 07:00:56 crc kubenswrapper[4559]: I1123 07:00:56.998948 4559 generic.go:334] "Generic (PLEG): container finished" podID="665d6fe5-5f01-48f1-bbe3-1ee83666e9d5" containerID="e306043b6c63645e4301c6d2ada6b062eb31ed1497ee6886297d2919099fddce" exitCode=0 Nov 23 07:00:56 crc kubenswrapper[4559]: I1123 07:00:56.998659 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5","Type":"ContainerDied","Data":"8bd770cea540459a48d5d916eefc70b323ab4befddb41271517ac7d3c56fd34e"} Nov 23 07:00:56 crc kubenswrapper[4559]: I1123 07:00:56.999001 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5","Type":"ContainerDied","Data":"8a0cc150042c2d08c3205b6303843fb4a3135a34fbf86beff1d50658f99f4754"} Nov 23 07:00:56 crc kubenswrapper[4559]: I1123 07:00:56.999018 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5","Type":"ContainerDied","Data":"e306043b6c63645e4301c6d2ada6b062eb31ed1497ee6886297d2919099fddce"} Nov 23 07:00:58 crc kubenswrapper[4559]: I1123 07:00:58.009329 4559 generic.go:334] "Generic (PLEG): container finished" podID="b08ced2f-75ec-4b10-9e87-e83572224054" containerID="dd9cc55d2d749115f38ba72ac2c037e554c026fa3b9dfb051ed05730645e4c3b" exitCode=0 Nov 23 07:00:58 crc kubenswrapper[4559]: I1123 07:00:58.009354 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7nhf7" event={"ID":"b08ced2f-75ec-4b10-9e87-e83572224054","Type":"ContainerDied","Data":"dd9cc55d2d749115f38ba72ac2c037e554c026fa3b9dfb051ed05730645e4c3b"} Nov 23 07:00:58 crc kubenswrapper[4559]: I1123 07:00:58.025484 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=7.232987219 podStartE2EDuration="11.025467178s" podCreationTimestamp="2025-11-23 07:00:47 +0000 UTC" firstStartedPulling="2025-11-23 07:00:51.736293013 +0000 UTC m=+953.758278627" lastFinishedPulling="2025-11-23 07:00:55.528772971 +0000 UTC m=+957.550758586" observedRunningTime="2025-11-23 07:00:56.020456502 +0000 UTC m=+958.042442116" watchObservedRunningTime="2025-11-23 07:00:58.025467178 +0000 UTC m=+960.047452792" Nov 23 07:00:59 crc kubenswrapper[4559]: I1123 07:00:59.310681 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7nhf7" Nov 23 07:00:59 crc kubenswrapper[4559]: I1123 07:00:59.481024 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b08ced2f-75ec-4b10-9e87-e83572224054-config-data\") pod \"b08ced2f-75ec-4b10-9e87-e83572224054\" (UID: \"b08ced2f-75ec-4b10-9e87-e83572224054\") " Nov 23 07:00:59 crc kubenswrapper[4559]: I1123 07:00:59.481158 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b08ced2f-75ec-4b10-9e87-e83572224054-scripts\") pod \"b08ced2f-75ec-4b10-9e87-e83572224054\" (UID: \"b08ced2f-75ec-4b10-9e87-e83572224054\") " Nov 23 07:00:59 crc kubenswrapper[4559]: I1123 07:00:59.481311 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b08ced2f-75ec-4b10-9e87-e83572224054-combined-ca-bundle\") pod \"b08ced2f-75ec-4b10-9e87-e83572224054\" (UID: \"b08ced2f-75ec-4b10-9e87-e83572224054\") " Nov 23 07:00:59 crc kubenswrapper[4559]: I1123 07:00:59.481344 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fg557\" (UniqueName: \"kubernetes.io/projected/b08ced2f-75ec-4b10-9e87-e83572224054-kube-api-access-fg557\") pod \"b08ced2f-75ec-4b10-9e87-e83572224054\" (UID: \"b08ced2f-75ec-4b10-9e87-e83572224054\") " Nov 23 07:00:59 crc kubenswrapper[4559]: I1123 07:00:59.486943 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b08ced2f-75ec-4b10-9e87-e83572224054-kube-api-access-fg557" (OuterVolumeSpecName: "kube-api-access-fg557") pod "b08ced2f-75ec-4b10-9e87-e83572224054" (UID: "b08ced2f-75ec-4b10-9e87-e83572224054"). InnerVolumeSpecName "kube-api-access-fg557". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:00:59 crc kubenswrapper[4559]: I1123 07:00:59.487795 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b08ced2f-75ec-4b10-9e87-e83572224054-scripts" (OuterVolumeSpecName: "scripts") pod "b08ced2f-75ec-4b10-9e87-e83572224054" (UID: "b08ced2f-75ec-4b10-9e87-e83572224054"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:59 crc kubenswrapper[4559]: I1123 07:00:59.511384 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b08ced2f-75ec-4b10-9e87-e83572224054-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b08ced2f-75ec-4b10-9e87-e83572224054" (UID: "b08ced2f-75ec-4b10-9e87-e83572224054"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:59 crc kubenswrapper[4559]: I1123 07:00:59.512421 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b08ced2f-75ec-4b10-9e87-e83572224054-config-data" (OuterVolumeSpecName: "config-data") pod "b08ced2f-75ec-4b10-9e87-e83572224054" (UID: "b08ced2f-75ec-4b10-9e87-e83572224054"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:00:59 crc kubenswrapper[4559]: I1123 07:00:59.584853 4559 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b08ced2f-75ec-4b10-9e87-e83572224054-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:59 crc kubenswrapper[4559]: I1123 07:00:59.584923 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b08ced2f-75ec-4b10-9e87-e83572224054-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:59 crc kubenswrapper[4559]: I1123 07:00:59.584942 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fg557\" (UniqueName: \"kubernetes.io/projected/b08ced2f-75ec-4b10-9e87-e83572224054-kube-api-access-fg557\") on node \"crc\" DevicePath \"\"" Nov 23 07:00:59 crc kubenswrapper[4559]: I1123 07:00:59.584954 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b08ced2f-75ec-4b10-9e87-e83572224054-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.026672 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7nhf7" event={"ID":"b08ced2f-75ec-4b10-9e87-e83572224054","Type":"ContainerDied","Data":"01def69877136d8d0f89c9f6864060b5f83389e44927a26f3a24e634aecf55f6"} Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.026988 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01def69877136d8d0f89c9f6864060b5f83389e44927a26f3a24e634aecf55f6" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.026731 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7nhf7" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.097554 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 23 07:01:00 crc kubenswrapper[4559]: E1123 07:01:00.098053 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b08ced2f-75ec-4b10-9e87-e83572224054" containerName="nova-cell0-conductor-db-sync" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.098070 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="b08ced2f-75ec-4b10-9e87-e83572224054" containerName="nova-cell0-conductor-db-sync" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.098313 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="b08ced2f-75ec-4b10-9e87-e83572224054" containerName="nova-cell0-conductor-db-sync" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.099073 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.102170 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-fhsg9" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.103437 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.106691 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.132944 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29398021-r88wd"] Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.134386 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29398021-r88wd" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.153016 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29398021-r88wd"] Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.200565 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlkgp\" (UniqueName: \"kubernetes.io/projected/da23948d-f4d4-42cc-9960-aaeb3efc2309-kube-api-access-rlkgp\") pod \"keystone-cron-29398021-r88wd\" (UID: \"da23948d-f4d4-42cc-9960-aaeb3efc2309\") " pod="openstack/keystone-cron-29398021-r88wd" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.200681 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da23948d-f4d4-42cc-9960-aaeb3efc2309-combined-ca-bundle\") pod \"keystone-cron-29398021-r88wd\" (UID: \"da23948d-f4d4-42cc-9960-aaeb3efc2309\") " pod="openstack/keystone-cron-29398021-r88wd" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.200736 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5666b61-0dcd-4fcd-a96a-44fa29cd3d84-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a5666b61-0dcd-4fcd-a96a-44fa29cd3d84\") " pod="openstack/nova-cell0-conductor-0" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.200763 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da23948d-f4d4-42cc-9960-aaeb3efc2309-config-data\") pod \"keystone-cron-29398021-r88wd\" (UID: \"da23948d-f4d4-42cc-9960-aaeb3efc2309\") " pod="openstack/keystone-cron-29398021-r88wd" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.200835 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/da23948d-f4d4-42cc-9960-aaeb3efc2309-fernet-keys\") pod \"keystone-cron-29398021-r88wd\" (UID: \"da23948d-f4d4-42cc-9960-aaeb3efc2309\") " pod="openstack/keystone-cron-29398021-r88wd" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.200878 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mx7z5\" (UniqueName: \"kubernetes.io/projected/a5666b61-0dcd-4fcd-a96a-44fa29cd3d84-kube-api-access-mx7z5\") pod \"nova-cell0-conductor-0\" (UID: \"a5666b61-0dcd-4fcd-a96a-44fa29cd3d84\") " pod="openstack/nova-cell0-conductor-0" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.200920 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5666b61-0dcd-4fcd-a96a-44fa29cd3d84-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a5666b61-0dcd-4fcd-a96a-44fa29cd3d84\") " pod="openstack/nova-cell0-conductor-0" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.302406 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da23948d-f4d4-42cc-9960-aaeb3efc2309-combined-ca-bundle\") pod \"keystone-cron-29398021-r88wd\" (UID: \"da23948d-f4d4-42cc-9960-aaeb3efc2309\") " pod="openstack/keystone-cron-29398021-r88wd" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.302456 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5666b61-0dcd-4fcd-a96a-44fa29cd3d84-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a5666b61-0dcd-4fcd-a96a-44fa29cd3d84\") " pod="openstack/nova-cell0-conductor-0" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.302488 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da23948d-f4d4-42cc-9960-aaeb3efc2309-config-data\") pod \"keystone-cron-29398021-r88wd\" (UID: \"da23948d-f4d4-42cc-9960-aaeb3efc2309\") " pod="openstack/keystone-cron-29398021-r88wd" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.302534 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/da23948d-f4d4-42cc-9960-aaeb3efc2309-fernet-keys\") pod \"keystone-cron-29398021-r88wd\" (UID: \"da23948d-f4d4-42cc-9960-aaeb3efc2309\") " pod="openstack/keystone-cron-29398021-r88wd" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.302551 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mx7z5\" (UniqueName: \"kubernetes.io/projected/a5666b61-0dcd-4fcd-a96a-44fa29cd3d84-kube-api-access-mx7z5\") pod \"nova-cell0-conductor-0\" (UID: \"a5666b61-0dcd-4fcd-a96a-44fa29cd3d84\") " pod="openstack/nova-cell0-conductor-0" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.303372 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5666b61-0dcd-4fcd-a96a-44fa29cd3d84-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a5666b61-0dcd-4fcd-a96a-44fa29cd3d84\") " pod="openstack/nova-cell0-conductor-0" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.303444 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlkgp\" (UniqueName: \"kubernetes.io/projected/da23948d-f4d4-42cc-9960-aaeb3efc2309-kube-api-access-rlkgp\") pod \"keystone-cron-29398021-r88wd\" (UID: \"da23948d-f4d4-42cc-9960-aaeb3efc2309\") " pod="openstack/keystone-cron-29398021-r88wd" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.307871 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5666b61-0dcd-4fcd-a96a-44fa29cd3d84-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a5666b61-0dcd-4fcd-a96a-44fa29cd3d84\") " pod="openstack/nova-cell0-conductor-0" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.308013 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da23948d-f4d4-42cc-9960-aaeb3efc2309-config-data\") pod \"keystone-cron-29398021-r88wd\" (UID: \"da23948d-f4d4-42cc-9960-aaeb3efc2309\") " pod="openstack/keystone-cron-29398021-r88wd" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.308276 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da23948d-f4d4-42cc-9960-aaeb3efc2309-combined-ca-bundle\") pod \"keystone-cron-29398021-r88wd\" (UID: \"da23948d-f4d4-42cc-9960-aaeb3efc2309\") " pod="openstack/keystone-cron-29398021-r88wd" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.314998 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5666b61-0dcd-4fcd-a96a-44fa29cd3d84-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a5666b61-0dcd-4fcd-a96a-44fa29cd3d84\") " pod="openstack/nova-cell0-conductor-0" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.316636 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/da23948d-f4d4-42cc-9960-aaeb3efc2309-fernet-keys\") pod \"keystone-cron-29398021-r88wd\" (UID: \"da23948d-f4d4-42cc-9960-aaeb3efc2309\") " pod="openstack/keystone-cron-29398021-r88wd" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.316880 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mx7z5\" (UniqueName: \"kubernetes.io/projected/a5666b61-0dcd-4fcd-a96a-44fa29cd3d84-kube-api-access-mx7z5\") pod \"nova-cell0-conductor-0\" (UID: \"a5666b61-0dcd-4fcd-a96a-44fa29cd3d84\") " pod="openstack/nova-cell0-conductor-0" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.329245 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlkgp\" (UniqueName: \"kubernetes.io/projected/da23948d-f4d4-42cc-9960-aaeb3efc2309-kube-api-access-rlkgp\") pod \"keystone-cron-29398021-r88wd\" (UID: \"da23948d-f4d4-42cc-9960-aaeb3efc2309\") " pod="openstack/keystone-cron-29398021-r88wd" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.413466 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.454772 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29398021-r88wd" Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.814215 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 23 07:01:00 crc kubenswrapper[4559]: W1123 07:01:00.839221 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5666b61_0dcd_4fcd_a96a_44fa29cd3d84.slice/crio-90ce1a2f60fadb8bde82ad34e85d463b043a10552034b30e99eb82a594d9de91 WatchSource:0}: Error finding container 90ce1a2f60fadb8bde82ad34e85d463b043a10552034b30e99eb82a594d9de91: Status 404 returned error can't find the container with id 90ce1a2f60fadb8bde82ad34e85d463b043a10552034b30e99eb82a594d9de91 Nov 23 07:01:00 crc kubenswrapper[4559]: I1123 07:01:00.879264 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29398021-r88wd"] Nov 23 07:01:00 crc kubenswrapper[4559]: W1123 07:01:00.880830 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda23948d_f4d4_42cc_9960_aaeb3efc2309.slice/crio-0252d1bdca1f1d2aa12d42a63f2ed1b1e79e2ec0ce831273e633186f037d70bf WatchSource:0}: Error finding container 0252d1bdca1f1d2aa12d42a63f2ed1b1e79e2ec0ce831273e633186f037d70bf: Status 404 returned error can't find the container with id 0252d1bdca1f1d2aa12d42a63f2ed1b1e79e2ec0ce831273e633186f037d70bf Nov 23 07:01:01 crc kubenswrapper[4559]: I1123 07:01:01.039285 4559 generic.go:334] "Generic (PLEG): container finished" podID="665d6fe5-5f01-48f1-bbe3-1ee83666e9d5" containerID="3430ff2eaf1f572f57a8701d797eb8c96b432ead2047c1b48c5fc67ecbc32953" exitCode=0 Nov 23 07:01:01 crc kubenswrapper[4559]: I1123 07:01:01.039381 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5","Type":"ContainerDied","Data":"3430ff2eaf1f572f57a8701d797eb8c96b432ead2047c1b48c5fc67ecbc32953"} Nov 23 07:01:01 crc kubenswrapper[4559]: I1123 07:01:01.041048 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a5666b61-0dcd-4fcd-a96a-44fa29cd3d84","Type":"ContainerStarted","Data":"90ce1a2f60fadb8bde82ad34e85d463b043a10552034b30e99eb82a594d9de91"} Nov 23 07:01:01 crc kubenswrapper[4559]: I1123 07:01:01.042387 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29398021-r88wd" event={"ID":"da23948d-f4d4-42cc-9960-aaeb3efc2309","Type":"ContainerStarted","Data":"0252d1bdca1f1d2aa12d42a63f2ed1b1e79e2ec0ce831273e633186f037d70bf"} Nov 23 07:01:02 crc kubenswrapper[4559]: I1123 07:01:02.711235 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:01:02 crc kubenswrapper[4559]: I1123 07:01:02.848087 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cn6cg\" (UniqueName: \"kubernetes.io/projected/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-kube-api-access-cn6cg\") pod \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\" (UID: \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\") " Nov 23 07:01:02 crc kubenswrapper[4559]: I1123 07:01:02.848252 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-combined-ca-bundle\") pod \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\" (UID: \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\") " Nov 23 07:01:02 crc kubenswrapper[4559]: I1123 07:01:02.848347 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-scripts\") pod \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\" (UID: \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\") " Nov 23 07:01:02 crc kubenswrapper[4559]: I1123 07:01:02.848425 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-sg-core-conf-yaml\") pod \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\" (UID: \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\") " Nov 23 07:01:02 crc kubenswrapper[4559]: I1123 07:01:02.848494 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-config-data\") pod \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\" (UID: \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\") " Nov 23 07:01:02 crc kubenswrapper[4559]: I1123 07:01:02.848619 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-run-httpd\") pod \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\" (UID: \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\") " Nov 23 07:01:02 crc kubenswrapper[4559]: I1123 07:01:02.848696 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-log-httpd\") pod \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\" (UID: \"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5\") " Nov 23 07:01:02 crc kubenswrapper[4559]: I1123 07:01:02.849108 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "665d6fe5-5f01-48f1-bbe3-1ee83666e9d5" (UID: "665d6fe5-5f01-48f1-bbe3-1ee83666e9d5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:01:02 crc kubenswrapper[4559]: I1123 07:01:02.849429 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "665d6fe5-5f01-48f1-bbe3-1ee83666e9d5" (UID: "665d6fe5-5f01-48f1-bbe3-1ee83666e9d5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:01:02 crc kubenswrapper[4559]: I1123 07:01:02.849459 4559 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:02 crc kubenswrapper[4559]: I1123 07:01:02.856586 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-kube-api-access-cn6cg" (OuterVolumeSpecName: "kube-api-access-cn6cg") pod "665d6fe5-5f01-48f1-bbe3-1ee83666e9d5" (UID: "665d6fe5-5f01-48f1-bbe3-1ee83666e9d5"). InnerVolumeSpecName "kube-api-access-cn6cg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:01:02 crc kubenswrapper[4559]: I1123 07:01:02.877338 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-scripts" (OuterVolumeSpecName: "scripts") pod "665d6fe5-5f01-48f1-bbe3-1ee83666e9d5" (UID: "665d6fe5-5f01-48f1-bbe3-1ee83666e9d5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:02 crc kubenswrapper[4559]: I1123 07:01:02.877359 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "665d6fe5-5f01-48f1-bbe3-1ee83666e9d5" (UID: "665d6fe5-5f01-48f1-bbe3-1ee83666e9d5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:02 crc kubenswrapper[4559]: I1123 07:01:02.906827 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "665d6fe5-5f01-48f1-bbe3-1ee83666e9d5" (UID: "665d6fe5-5f01-48f1-bbe3-1ee83666e9d5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:02 crc kubenswrapper[4559]: I1123 07:01:02.923055 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-config-data" (OuterVolumeSpecName: "config-data") pod "665d6fe5-5f01-48f1-bbe3-1ee83666e9d5" (UID: "665d6fe5-5f01-48f1-bbe3-1ee83666e9d5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:02 crc kubenswrapper[4559]: I1123 07:01:02.951926 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:02 crc kubenswrapper[4559]: I1123 07:01:02.951969 4559 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:02 crc kubenswrapper[4559]: I1123 07:01:02.951983 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cn6cg\" (UniqueName: \"kubernetes.io/projected/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-kube-api-access-cn6cg\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:02 crc kubenswrapper[4559]: I1123 07:01:02.951999 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:02 crc kubenswrapper[4559]: I1123 07:01:02.952009 4559 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:02 crc kubenswrapper[4559]: I1123 07:01:02.952020 4559 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.067941 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a5666b61-0dcd-4fcd-a96a-44fa29cd3d84","Type":"ContainerStarted","Data":"0b22eff2584d469f47433b92944cefbc48b550caff9ff58d6a8c6741634b1aa0"} Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.068125 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.069145 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29398021-r88wd" event={"ID":"da23948d-f4d4-42cc-9960-aaeb3efc2309","Type":"ContainerStarted","Data":"1cae78a76e7d5cd7c36b8bd2c24837aa3c8b158061381bc36704b198bf8236e4"} Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.073262 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"665d6fe5-5f01-48f1-bbe3-1ee83666e9d5","Type":"ContainerDied","Data":"213f547a280bfd1fcd1151aee3b760e8c7deaafc6f594ce952ad900811276990"} Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.073307 4559 scope.go:117] "RemoveContainer" containerID="8bd770cea540459a48d5d916eefc70b323ab4befddb41271517ac7d3c56fd34e" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.073349 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.089110 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=3.08909378 podStartE2EDuration="3.08909378s" podCreationTimestamp="2025-11-23 07:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:01:03.087181093 +0000 UTC m=+965.109166707" watchObservedRunningTime="2025-11-23 07:01:03.08909378 +0000 UTC m=+965.111079394" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.101710 4559 scope.go:117] "RemoveContainer" containerID="8a0cc150042c2d08c3205b6303843fb4a3135a34fbf86beff1d50658f99f4754" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.104912 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29398021-r88wd" podStartSLOduration=3.104898495 podStartE2EDuration="3.104898495s" podCreationTimestamp="2025-11-23 07:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:01:03.102310406 +0000 UTC m=+965.124296030" watchObservedRunningTime="2025-11-23 07:01:03.104898495 +0000 UTC m=+965.126884109" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.126025 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.136773 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.137392 4559 scope.go:117] "RemoveContainer" containerID="e306043b6c63645e4301c6d2ada6b062eb31ed1497ee6886297d2919099fddce" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.145239 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:01:03 crc kubenswrapper[4559]: E1123 07:01:03.145705 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="665d6fe5-5f01-48f1-bbe3-1ee83666e9d5" containerName="ceilometer-notification-agent" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.145727 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="665d6fe5-5f01-48f1-bbe3-1ee83666e9d5" containerName="ceilometer-notification-agent" Nov 23 07:01:03 crc kubenswrapper[4559]: E1123 07:01:03.145753 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="665d6fe5-5f01-48f1-bbe3-1ee83666e9d5" containerName="ceilometer-central-agent" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.145760 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="665d6fe5-5f01-48f1-bbe3-1ee83666e9d5" containerName="ceilometer-central-agent" Nov 23 07:01:03 crc kubenswrapper[4559]: E1123 07:01:03.145775 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="665d6fe5-5f01-48f1-bbe3-1ee83666e9d5" containerName="sg-core" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.145783 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="665d6fe5-5f01-48f1-bbe3-1ee83666e9d5" containerName="sg-core" Nov 23 07:01:03 crc kubenswrapper[4559]: E1123 07:01:03.145792 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="665d6fe5-5f01-48f1-bbe3-1ee83666e9d5" containerName="proxy-httpd" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.145798 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="665d6fe5-5f01-48f1-bbe3-1ee83666e9d5" containerName="proxy-httpd" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.145982 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="665d6fe5-5f01-48f1-bbe3-1ee83666e9d5" containerName="sg-core" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.146002 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="665d6fe5-5f01-48f1-bbe3-1ee83666e9d5" containerName="ceilometer-notification-agent" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.146020 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="665d6fe5-5f01-48f1-bbe3-1ee83666e9d5" containerName="proxy-httpd" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.146029 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="665d6fe5-5f01-48f1-bbe3-1ee83666e9d5" containerName="ceilometer-central-agent" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.147623 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.157108 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.157703 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.157881 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.164713 4559 scope.go:117] "RemoveContainer" containerID="3430ff2eaf1f572f57a8701d797eb8c96b432ead2047c1b48c5fc67ecbc32953" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.261346 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-scripts\") pod \"ceilometer-0\" (UID: \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\") " pod="openstack/ceilometer-0" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.261602 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-config-data\") pod \"ceilometer-0\" (UID: \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\") " pod="openstack/ceilometer-0" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.261858 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-run-httpd\") pod \"ceilometer-0\" (UID: \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\") " pod="openstack/ceilometer-0" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.262157 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glgpq\" (UniqueName: \"kubernetes.io/projected/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-kube-api-access-glgpq\") pod \"ceilometer-0\" (UID: \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\") " pod="openstack/ceilometer-0" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.262212 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\") " pod="openstack/ceilometer-0" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.262294 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-log-httpd\") pod \"ceilometer-0\" (UID: \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\") " pod="openstack/ceilometer-0" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.262321 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\") " pod="openstack/ceilometer-0" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.365413 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-log-httpd\") pod \"ceilometer-0\" (UID: \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\") " pod="openstack/ceilometer-0" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.365462 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\") " pod="openstack/ceilometer-0" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.365530 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-scripts\") pod \"ceilometer-0\" (UID: \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\") " pod="openstack/ceilometer-0" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.365573 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-config-data\") pod \"ceilometer-0\" (UID: \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\") " pod="openstack/ceilometer-0" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.365595 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-run-httpd\") pod \"ceilometer-0\" (UID: \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\") " pod="openstack/ceilometer-0" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.365668 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glgpq\" (UniqueName: \"kubernetes.io/projected/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-kube-api-access-glgpq\") pod \"ceilometer-0\" (UID: \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\") " pod="openstack/ceilometer-0" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.365694 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\") " pod="openstack/ceilometer-0" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.366617 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-run-httpd\") pod \"ceilometer-0\" (UID: \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\") " pod="openstack/ceilometer-0" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.367663 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-log-httpd\") pod \"ceilometer-0\" (UID: \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\") " pod="openstack/ceilometer-0" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.369083 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\") " pod="openstack/ceilometer-0" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.369620 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\") " pod="openstack/ceilometer-0" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.369944 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-scripts\") pod \"ceilometer-0\" (UID: \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\") " pod="openstack/ceilometer-0" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.372385 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-config-data\") pod \"ceilometer-0\" (UID: \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\") " pod="openstack/ceilometer-0" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.382287 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glgpq\" (UniqueName: \"kubernetes.io/projected/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-kube-api-access-glgpq\") pod \"ceilometer-0\" (UID: \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\") " pod="openstack/ceilometer-0" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.476743 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:01:03 crc kubenswrapper[4559]: I1123 07:01:03.867290 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:01:03 crc kubenswrapper[4559]: W1123 07:01:03.871586 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode76b4aa1_64f3_480b_a134_5ac6f6c02f04.slice/crio-3d93e1f9bbb249f41a971a795b9e9ee7c19b6e1921e5be767ce24de7d4830925 WatchSource:0}: Error finding container 3d93e1f9bbb249f41a971a795b9e9ee7c19b6e1921e5be767ce24de7d4830925: Status 404 returned error can't find the container with id 3d93e1f9bbb249f41a971a795b9e9ee7c19b6e1921e5be767ce24de7d4830925 Nov 23 07:01:04 crc kubenswrapper[4559]: I1123 07:01:04.086575 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e76b4aa1-64f3-480b-a134-5ac6f6c02f04","Type":"ContainerStarted","Data":"3d93e1f9bbb249f41a971a795b9e9ee7c19b6e1921e5be767ce24de7d4830925"} Nov 23 07:01:04 crc kubenswrapper[4559]: I1123 07:01:04.285750 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="665d6fe5-5f01-48f1-bbe3-1ee83666e9d5" path="/var/lib/kubelet/pods/665d6fe5-5f01-48f1-bbe3-1ee83666e9d5/volumes" Nov 23 07:01:05 crc kubenswrapper[4559]: I1123 07:01:05.113602 4559 generic.go:334] "Generic (PLEG): container finished" podID="da23948d-f4d4-42cc-9960-aaeb3efc2309" containerID="1cae78a76e7d5cd7c36b8bd2c24837aa3c8b158061381bc36704b198bf8236e4" exitCode=0 Nov 23 07:01:05 crc kubenswrapper[4559]: I1123 07:01:05.113687 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29398021-r88wd" event={"ID":"da23948d-f4d4-42cc-9960-aaeb3efc2309","Type":"ContainerDied","Data":"1cae78a76e7d5cd7c36b8bd2c24837aa3c8b158061381bc36704b198bf8236e4"} Nov 23 07:01:05 crc kubenswrapper[4559]: I1123 07:01:05.118051 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e76b4aa1-64f3-480b-a134-5ac6f6c02f04","Type":"ContainerStarted","Data":"54e5ac38d914c27d936b52c687e86987b540816ad3e7956c3f4263888aac90b4"} Nov 23 07:01:06 crc kubenswrapper[4559]: I1123 07:01:06.131168 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e76b4aa1-64f3-480b-a134-5ac6f6c02f04","Type":"ContainerStarted","Data":"85e3166a92990c3ebc8bb481e3392e6bcd5139c8f8518e7ca870262593a5b5ef"} Nov 23 07:01:06 crc kubenswrapper[4559]: I1123 07:01:06.419240 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29398021-r88wd" Nov 23 07:01:06 crc kubenswrapper[4559]: I1123 07:01:06.541230 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da23948d-f4d4-42cc-9960-aaeb3efc2309-combined-ca-bundle\") pod \"da23948d-f4d4-42cc-9960-aaeb3efc2309\" (UID: \"da23948d-f4d4-42cc-9960-aaeb3efc2309\") " Nov 23 07:01:06 crc kubenswrapper[4559]: I1123 07:01:06.541341 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/da23948d-f4d4-42cc-9960-aaeb3efc2309-fernet-keys\") pod \"da23948d-f4d4-42cc-9960-aaeb3efc2309\" (UID: \"da23948d-f4d4-42cc-9960-aaeb3efc2309\") " Nov 23 07:01:06 crc kubenswrapper[4559]: I1123 07:01:06.541407 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlkgp\" (UniqueName: \"kubernetes.io/projected/da23948d-f4d4-42cc-9960-aaeb3efc2309-kube-api-access-rlkgp\") pod \"da23948d-f4d4-42cc-9960-aaeb3efc2309\" (UID: \"da23948d-f4d4-42cc-9960-aaeb3efc2309\") " Nov 23 07:01:06 crc kubenswrapper[4559]: I1123 07:01:06.541512 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da23948d-f4d4-42cc-9960-aaeb3efc2309-config-data\") pod \"da23948d-f4d4-42cc-9960-aaeb3efc2309\" (UID: \"da23948d-f4d4-42cc-9960-aaeb3efc2309\") " Nov 23 07:01:06 crc kubenswrapper[4559]: I1123 07:01:06.548412 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da23948d-f4d4-42cc-9960-aaeb3efc2309-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "da23948d-f4d4-42cc-9960-aaeb3efc2309" (UID: "da23948d-f4d4-42cc-9960-aaeb3efc2309"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:06 crc kubenswrapper[4559]: I1123 07:01:06.548470 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da23948d-f4d4-42cc-9960-aaeb3efc2309-kube-api-access-rlkgp" (OuterVolumeSpecName: "kube-api-access-rlkgp") pod "da23948d-f4d4-42cc-9960-aaeb3efc2309" (UID: "da23948d-f4d4-42cc-9960-aaeb3efc2309"). InnerVolumeSpecName "kube-api-access-rlkgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:01:06 crc kubenswrapper[4559]: I1123 07:01:06.570290 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da23948d-f4d4-42cc-9960-aaeb3efc2309-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "da23948d-f4d4-42cc-9960-aaeb3efc2309" (UID: "da23948d-f4d4-42cc-9960-aaeb3efc2309"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:06 crc kubenswrapper[4559]: I1123 07:01:06.588154 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da23948d-f4d4-42cc-9960-aaeb3efc2309-config-data" (OuterVolumeSpecName: "config-data") pod "da23948d-f4d4-42cc-9960-aaeb3efc2309" (UID: "da23948d-f4d4-42cc-9960-aaeb3efc2309"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:06 crc kubenswrapper[4559]: I1123 07:01:06.642803 4559 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/da23948d-f4d4-42cc-9960-aaeb3efc2309-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:06 crc kubenswrapper[4559]: I1123 07:01:06.643032 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlkgp\" (UniqueName: \"kubernetes.io/projected/da23948d-f4d4-42cc-9960-aaeb3efc2309-kube-api-access-rlkgp\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:06 crc kubenswrapper[4559]: I1123 07:01:06.643043 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da23948d-f4d4-42cc-9960-aaeb3efc2309-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:06 crc kubenswrapper[4559]: I1123 07:01:06.643051 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da23948d-f4d4-42cc-9960-aaeb3efc2309-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:07 crc kubenswrapper[4559]: I1123 07:01:07.150208 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e76b4aa1-64f3-480b-a134-5ac6f6c02f04","Type":"ContainerStarted","Data":"4b08563b50b2eb7ce1a5a2e63a6d28fef78d8987d89051f15a197f44fe01d399"} Nov 23 07:01:07 crc kubenswrapper[4559]: I1123 07:01:07.153147 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29398021-r88wd" Nov 23 07:01:07 crc kubenswrapper[4559]: I1123 07:01:07.153113 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29398021-r88wd" event={"ID":"da23948d-f4d4-42cc-9960-aaeb3efc2309","Type":"ContainerDied","Data":"0252d1bdca1f1d2aa12d42a63f2ed1b1e79e2ec0ce831273e633186f037d70bf"} Nov 23 07:01:07 crc kubenswrapper[4559]: I1123 07:01:07.153233 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0252d1bdca1f1d2aa12d42a63f2ed1b1e79e2ec0ce831273e633186f037d70bf" Nov 23 07:01:08 crc kubenswrapper[4559]: I1123 07:01:08.163333 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e76b4aa1-64f3-480b-a134-5ac6f6c02f04","Type":"ContainerStarted","Data":"f9044311486a20b6b19f917e0121c85bad3325fd50152023e39c848cc28da22d"} Nov 23 07:01:08 crc kubenswrapper[4559]: I1123 07:01:08.163722 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 23 07:01:08 crc kubenswrapper[4559]: I1123 07:01:08.190262 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.422281235 podStartE2EDuration="5.190233634s" podCreationTimestamp="2025-11-23 07:01:03 +0000 UTC" firstStartedPulling="2025-11-23 07:01:03.874522082 +0000 UTC m=+965.896507696" lastFinishedPulling="2025-11-23 07:01:07.642474482 +0000 UTC m=+969.664460095" observedRunningTime="2025-11-23 07:01:08.17976736 +0000 UTC m=+970.201752974" watchObservedRunningTime="2025-11-23 07:01:08.190233634 +0000 UTC m=+970.212219248" Nov 23 07:01:10 crc kubenswrapper[4559]: I1123 07:01:10.435693 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 23 07:01:10 crc kubenswrapper[4559]: I1123 07:01:10.841777 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-vfsc8"] Nov 23 07:01:10 crc kubenswrapper[4559]: E1123 07:01:10.842177 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da23948d-f4d4-42cc-9960-aaeb3efc2309" containerName="keystone-cron" Nov 23 07:01:10 crc kubenswrapper[4559]: I1123 07:01:10.842194 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="da23948d-f4d4-42cc-9960-aaeb3efc2309" containerName="keystone-cron" Nov 23 07:01:10 crc kubenswrapper[4559]: I1123 07:01:10.842367 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="da23948d-f4d4-42cc-9960-aaeb3efc2309" containerName="keystone-cron" Nov 23 07:01:10 crc kubenswrapper[4559]: I1123 07:01:10.842929 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vfsc8" Nov 23 07:01:10 crc kubenswrapper[4559]: I1123 07:01:10.844901 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 23 07:01:10 crc kubenswrapper[4559]: I1123 07:01:10.845947 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 23 07:01:10 crc kubenswrapper[4559]: I1123 07:01:10.849177 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-vfsc8"] Nov 23 07:01:10 crc kubenswrapper[4559]: I1123 07:01:10.969481 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 23 07:01:10 crc kubenswrapper[4559]: I1123 07:01:10.970993 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:01:10 crc kubenswrapper[4559]: I1123 07:01:10.972763 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.006699 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.032082 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baf469bd-b908-4271-8688-ee06c3b64e0b-config-data\") pod \"nova-cell0-cell-mapping-vfsc8\" (UID: \"baf469bd-b908-4271-8688-ee06c3b64e0b\") " pod="openstack/nova-cell0-cell-mapping-vfsc8" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.032187 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfhlg\" (UniqueName: \"kubernetes.io/projected/baf469bd-b908-4271-8688-ee06c3b64e0b-kube-api-access-lfhlg\") pod \"nova-cell0-cell-mapping-vfsc8\" (UID: \"baf469bd-b908-4271-8688-ee06c3b64e0b\") " pod="openstack/nova-cell0-cell-mapping-vfsc8" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.032298 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baf469bd-b908-4271-8688-ee06c3b64e0b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vfsc8\" (UID: \"baf469bd-b908-4271-8688-ee06c3b64e0b\") " pod="openstack/nova-cell0-cell-mapping-vfsc8" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.032359 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baf469bd-b908-4271-8688-ee06c3b64e0b-scripts\") pod \"nova-cell0-cell-mapping-vfsc8\" (UID: \"baf469bd-b908-4271-8688-ee06c3b64e0b\") " pod="openstack/nova-cell0-cell-mapping-vfsc8" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.052800 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.054335 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.058780 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.064391 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.133911 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.134766 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a05da27-e06e-4f41-91ae-ad450a604acb-config-data\") pod \"nova-api-0\" (UID: \"0a05da27-e06e-4f41-91ae-ad450a604acb\") " pod="openstack/nova-api-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.134956 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.135174 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baf469bd-b908-4271-8688-ee06c3b64e0b-scripts\") pod \"nova-cell0-cell-mapping-vfsc8\" (UID: \"baf469bd-b908-4271-8688-ee06c3b64e0b\") " pod="openstack/nova-cell0-cell-mapping-vfsc8" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.135257 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a05da27-e06e-4f41-91ae-ad450a604acb-logs\") pod \"nova-api-0\" (UID: \"0a05da27-e06e-4f41-91ae-ad450a604acb\") " pod="openstack/nova-api-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.135379 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baf469bd-b908-4271-8688-ee06c3b64e0b-config-data\") pod \"nova-cell0-cell-mapping-vfsc8\" (UID: \"baf469bd-b908-4271-8688-ee06c3b64e0b\") " pod="openstack/nova-cell0-cell-mapping-vfsc8" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.135518 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d92qt\" (UniqueName: \"kubernetes.io/projected/0a05da27-e06e-4f41-91ae-ad450a604acb-kube-api-access-d92qt\") pod \"nova-api-0\" (UID: \"0a05da27-e06e-4f41-91ae-ad450a604acb\") " pod="openstack/nova-api-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.135591 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfhlg\" (UniqueName: \"kubernetes.io/projected/baf469bd-b908-4271-8688-ee06c3b64e0b-kube-api-access-lfhlg\") pod \"nova-cell0-cell-mapping-vfsc8\" (UID: \"baf469bd-b908-4271-8688-ee06c3b64e0b\") " pod="openstack/nova-cell0-cell-mapping-vfsc8" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.135663 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a05da27-e06e-4f41-91ae-ad450a604acb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0a05da27-e06e-4f41-91ae-ad450a604acb\") " pod="openstack/nova-api-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.135810 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baf469bd-b908-4271-8688-ee06c3b64e0b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vfsc8\" (UID: \"baf469bd-b908-4271-8688-ee06c3b64e0b\") " pod="openstack/nova-cell0-cell-mapping-vfsc8" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.140061 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.144569 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baf469bd-b908-4271-8688-ee06c3b64e0b-config-data\") pod \"nova-cell0-cell-mapping-vfsc8\" (UID: \"baf469bd-b908-4271-8688-ee06c3b64e0b\") " pod="openstack/nova-cell0-cell-mapping-vfsc8" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.146434 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baf469bd-b908-4271-8688-ee06c3b64e0b-scripts\") pod \"nova-cell0-cell-mapping-vfsc8\" (UID: \"baf469bd-b908-4271-8688-ee06c3b64e0b\") " pod="openstack/nova-cell0-cell-mapping-vfsc8" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.154306 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baf469bd-b908-4271-8688-ee06c3b64e0b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vfsc8\" (UID: \"baf469bd-b908-4271-8688-ee06c3b64e0b\") " pod="openstack/nova-cell0-cell-mapping-vfsc8" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.167191 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.167614 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfhlg\" (UniqueName: \"kubernetes.io/projected/baf469bd-b908-4271-8688-ee06c3b64e0b-kube-api-access-lfhlg\") pod \"nova-cell0-cell-mapping-vfsc8\" (UID: \"baf469bd-b908-4271-8688-ee06c3b64e0b\") " pod="openstack/nova-cell0-cell-mapping-vfsc8" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.179096 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77cbccc799-6bxp4"] Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.180597 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.187820 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.189382 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.194048 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.204602 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77cbccc799-6bxp4"] Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.257901 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d92qt\" (UniqueName: \"kubernetes.io/projected/0a05da27-e06e-4f41-91ae-ad450a604acb-kube-api-access-d92qt\") pod \"nova-api-0\" (UID: \"0a05da27-e06e-4f41-91ae-ad450a604acb\") " pod="openstack/nova-api-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.257996 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a05da27-e06e-4f41-91ae-ad450a604acb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0a05da27-e06e-4f41-91ae-ad450a604acb\") " pod="openstack/nova-api-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.258129 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5\") " pod="openstack/nova-metadata-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.258185 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfl89\" (UniqueName: \"kubernetes.io/projected/369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5-kube-api-access-rfl89\") pod \"nova-metadata-0\" (UID: \"369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5\") " pod="openstack/nova-metadata-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.258420 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a05da27-e06e-4f41-91ae-ad450a604acb-config-data\") pod \"nova-api-0\" (UID: \"0a05da27-e06e-4f41-91ae-ad450a604acb\") " pod="openstack/nova-api-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.258470 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5-logs\") pod \"nova-metadata-0\" (UID: \"369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5\") " pod="openstack/nova-metadata-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.258499 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a05da27-e06e-4f41-91ae-ad450a604acb-logs\") pod \"nova-api-0\" (UID: \"0a05da27-e06e-4f41-91ae-ad450a604acb\") " pod="openstack/nova-api-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.258758 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5-config-data\") pod \"nova-metadata-0\" (UID: \"369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5\") " pod="openstack/nova-metadata-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.264327 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a05da27-e06e-4f41-91ae-ad450a604acb-config-data\") pod \"nova-api-0\" (UID: \"0a05da27-e06e-4f41-91ae-ad450a604acb\") " pod="openstack/nova-api-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.264932 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a05da27-e06e-4f41-91ae-ad450a604acb-logs\") pod \"nova-api-0\" (UID: \"0a05da27-e06e-4f41-91ae-ad450a604acb\") " pod="openstack/nova-api-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.265461 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a05da27-e06e-4f41-91ae-ad450a604acb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0a05da27-e06e-4f41-91ae-ad450a604acb\") " pod="openstack/nova-api-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.272319 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.288435 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d92qt\" (UniqueName: \"kubernetes.io/projected/0a05da27-e06e-4f41-91ae-ad450a604acb-kube-api-access-d92qt\") pod \"nova-api-0\" (UID: \"0a05da27-e06e-4f41-91ae-ad450a604acb\") " pod="openstack/nova-api-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.341699 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-99bwv"] Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.343742 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-99bwv" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.344324 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-99bwv"] Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.370788 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-ovsdbserver-nb\") pod \"dnsmasq-dns-77cbccc799-6bxp4\" (UID: \"9e626278-ea64-4839-8cf2-852c7d336dbb\") " pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.370838 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2522249b-6744-4b19-9ffd-d5dbad83a1bd-config-data\") pod \"nova-scheduler-0\" (UID: \"2522249b-6744-4b19-9ffd-d5dbad83a1bd\") " pod="openstack/nova-scheduler-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.370926 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-config\") pod \"dnsmasq-dns-77cbccc799-6bxp4\" (UID: \"9e626278-ea64-4839-8cf2-852c7d336dbb\") " pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.370951 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87pcm\" (UniqueName: \"kubernetes.io/projected/2522249b-6744-4b19-9ffd-d5dbad83a1bd-kube-api-access-87pcm\") pod \"nova-scheduler-0\" (UID: \"2522249b-6744-4b19-9ffd-d5dbad83a1bd\") " pod="openstack/nova-scheduler-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.370980 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8n6zn\" (UniqueName: \"kubernetes.io/projected/9e626278-ea64-4839-8cf2-852c7d336dbb-kube-api-access-8n6zn\") pod \"dnsmasq-dns-77cbccc799-6bxp4\" (UID: \"9e626278-ea64-4839-8cf2-852c7d336dbb\") " pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.371040 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5\") " pod="openstack/nova-metadata-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.376111 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-dns-svc\") pod \"dnsmasq-dns-77cbccc799-6bxp4\" (UID: \"9e626278-ea64-4839-8cf2-852c7d336dbb\") " pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.376556 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb7a9144-0988-44cc-9596-0cdf01b5abb3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bb7a9144-0988-44cc-9596-0cdf01b5abb3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.376608 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-dns-swift-storage-0\") pod \"dnsmasq-dns-77cbccc799-6bxp4\" (UID: \"9e626278-ea64-4839-8cf2-852c7d336dbb\") " pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.376668 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfl89\" (UniqueName: \"kubernetes.io/projected/369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5-kube-api-access-rfl89\") pod \"nova-metadata-0\" (UID: \"369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5\") " pod="openstack/nova-metadata-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.376692 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlcgm\" (UniqueName: \"kubernetes.io/projected/bb7a9144-0988-44cc-9596-0cdf01b5abb3-kube-api-access-wlcgm\") pod \"nova-cell1-novncproxy-0\" (UID: \"bb7a9144-0988-44cc-9596-0cdf01b5abb3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.376801 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-ovsdbserver-sb\") pod \"dnsmasq-dns-77cbccc799-6bxp4\" (UID: \"9e626278-ea64-4839-8cf2-852c7d336dbb\") " pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.376852 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5-logs\") pod \"nova-metadata-0\" (UID: \"369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5\") " pod="openstack/nova-metadata-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.376902 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7a9144-0988-44cc-9596-0cdf01b5abb3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bb7a9144-0988-44cc-9596-0cdf01b5abb3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.376925 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5-config-data\") pod \"nova-metadata-0\" (UID: \"369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5\") " pod="openstack/nova-metadata-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.376972 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2522249b-6744-4b19-9ffd-d5dbad83a1bd-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2522249b-6744-4b19-9ffd-d5dbad83a1bd\") " pod="openstack/nova-scheduler-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.377544 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5-logs\") pod \"nova-metadata-0\" (UID: \"369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5\") " pod="openstack/nova-metadata-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.378869 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5\") " pod="openstack/nova-metadata-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.385881 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5-config-data\") pod \"nova-metadata-0\" (UID: \"369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5\") " pod="openstack/nova-metadata-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.391513 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfl89\" (UniqueName: \"kubernetes.io/projected/369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5-kube-api-access-rfl89\") pod \"nova-metadata-0\" (UID: \"369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5\") " pod="openstack/nova-metadata-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.465075 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vfsc8" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.478558 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2522249b-6744-4b19-9ffd-d5dbad83a1bd-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2522249b-6744-4b19-9ffd-d5dbad83a1bd\") " pod="openstack/nova-scheduler-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.478599 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-ovsdbserver-nb\") pod \"dnsmasq-dns-77cbccc799-6bxp4\" (UID: \"9e626278-ea64-4839-8cf2-852c7d336dbb\") " pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.478622 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2522249b-6744-4b19-9ffd-d5dbad83a1bd-config-data\") pod \"nova-scheduler-0\" (UID: \"2522249b-6744-4b19-9ffd-d5dbad83a1bd\") " pod="openstack/nova-scheduler-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.478700 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-config\") pod \"dnsmasq-dns-77cbccc799-6bxp4\" (UID: \"9e626278-ea64-4839-8cf2-852c7d336dbb\") " pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.478721 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87pcm\" (UniqueName: \"kubernetes.io/projected/2522249b-6744-4b19-9ffd-d5dbad83a1bd-kube-api-access-87pcm\") pod \"nova-scheduler-0\" (UID: \"2522249b-6744-4b19-9ffd-d5dbad83a1bd\") " pod="openstack/nova-scheduler-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.478739 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8n6zn\" (UniqueName: \"kubernetes.io/projected/9e626278-ea64-4839-8cf2-852c7d336dbb-kube-api-access-8n6zn\") pod \"dnsmasq-dns-77cbccc799-6bxp4\" (UID: \"9e626278-ea64-4839-8cf2-852c7d336dbb\") " pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.478774 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-dns-svc\") pod \"dnsmasq-dns-77cbccc799-6bxp4\" (UID: \"9e626278-ea64-4839-8cf2-852c7d336dbb\") " pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.478794 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb7a9144-0988-44cc-9596-0cdf01b5abb3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bb7a9144-0988-44cc-9596-0cdf01b5abb3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.478813 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-dns-swift-storage-0\") pod \"dnsmasq-dns-77cbccc799-6bxp4\" (UID: \"9e626278-ea64-4839-8cf2-852c7d336dbb\") " pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.478834 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgctz\" (UniqueName: \"kubernetes.io/projected/63adda33-68e6-456d-beac-04903b136067-kube-api-access-jgctz\") pod \"redhat-operators-99bwv\" (UID: \"63adda33-68e6-456d-beac-04903b136067\") " pod="openshift-marketplace/redhat-operators-99bwv" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.478853 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlcgm\" (UniqueName: \"kubernetes.io/projected/bb7a9144-0988-44cc-9596-0cdf01b5abb3-kube-api-access-wlcgm\") pod \"nova-cell1-novncproxy-0\" (UID: \"bb7a9144-0988-44cc-9596-0cdf01b5abb3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.478872 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63adda33-68e6-456d-beac-04903b136067-utilities\") pod \"redhat-operators-99bwv\" (UID: \"63adda33-68e6-456d-beac-04903b136067\") " pod="openshift-marketplace/redhat-operators-99bwv" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.478904 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-ovsdbserver-sb\") pod \"dnsmasq-dns-77cbccc799-6bxp4\" (UID: \"9e626278-ea64-4839-8cf2-852c7d336dbb\") " pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.479634 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-dns-svc\") pod \"dnsmasq-dns-77cbccc799-6bxp4\" (UID: \"9e626278-ea64-4839-8cf2-852c7d336dbb\") " pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.479779 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-config\") pod \"dnsmasq-dns-77cbccc799-6bxp4\" (UID: \"9e626278-ea64-4839-8cf2-852c7d336dbb\") " pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.479919 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-ovsdbserver-nb\") pod \"dnsmasq-dns-77cbccc799-6bxp4\" (UID: \"9e626278-ea64-4839-8cf2-852c7d336dbb\") " pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.478918 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63adda33-68e6-456d-beac-04903b136067-catalog-content\") pod \"redhat-operators-99bwv\" (UID: \"63adda33-68e6-456d-beac-04903b136067\") " pod="openshift-marketplace/redhat-operators-99bwv" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.480174 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7a9144-0988-44cc-9596-0cdf01b5abb3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bb7a9144-0988-44cc-9596-0cdf01b5abb3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.481867 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2522249b-6744-4b19-9ffd-d5dbad83a1bd-config-data\") pod \"nova-scheduler-0\" (UID: \"2522249b-6744-4b19-9ffd-d5dbad83a1bd\") " pod="openstack/nova-scheduler-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.483056 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-dns-swift-storage-0\") pod \"dnsmasq-dns-77cbccc799-6bxp4\" (UID: \"9e626278-ea64-4839-8cf2-852c7d336dbb\") " pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.483134 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-ovsdbserver-sb\") pod \"dnsmasq-dns-77cbccc799-6bxp4\" (UID: \"9e626278-ea64-4839-8cf2-852c7d336dbb\") " pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.483995 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2522249b-6744-4b19-9ffd-d5dbad83a1bd-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2522249b-6744-4b19-9ffd-d5dbad83a1bd\") " pod="openstack/nova-scheduler-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.489415 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7a9144-0988-44cc-9596-0cdf01b5abb3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bb7a9144-0988-44cc-9596-0cdf01b5abb3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.494147 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb7a9144-0988-44cc-9596-0cdf01b5abb3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bb7a9144-0988-44cc-9596-0cdf01b5abb3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.494299 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87pcm\" (UniqueName: \"kubernetes.io/projected/2522249b-6744-4b19-9ffd-d5dbad83a1bd-kube-api-access-87pcm\") pod \"nova-scheduler-0\" (UID: \"2522249b-6744-4b19-9ffd-d5dbad83a1bd\") " pod="openstack/nova-scheduler-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.495160 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8n6zn\" (UniqueName: \"kubernetes.io/projected/9e626278-ea64-4839-8cf2-852c7d336dbb-kube-api-access-8n6zn\") pod \"dnsmasq-dns-77cbccc799-6bxp4\" (UID: \"9e626278-ea64-4839-8cf2-852c7d336dbb\") " pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.495727 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlcgm\" (UniqueName: \"kubernetes.io/projected/bb7a9144-0988-44cc-9596-0cdf01b5abb3-kube-api-access-wlcgm\") pod \"nova-cell1-novncproxy-0\" (UID: \"bb7a9144-0988-44cc-9596-0cdf01b5abb3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.544494 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.567697 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.581783 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgctz\" (UniqueName: \"kubernetes.io/projected/63adda33-68e6-456d-beac-04903b136067-kube-api-access-jgctz\") pod \"redhat-operators-99bwv\" (UID: \"63adda33-68e6-456d-beac-04903b136067\") " pod="openshift-marketplace/redhat-operators-99bwv" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.581843 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63adda33-68e6-456d-beac-04903b136067-utilities\") pod \"redhat-operators-99bwv\" (UID: \"63adda33-68e6-456d-beac-04903b136067\") " pod="openshift-marketplace/redhat-operators-99bwv" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.581898 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63adda33-68e6-456d-beac-04903b136067-catalog-content\") pod \"redhat-operators-99bwv\" (UID: \"63adda33-68e6-456d-beac-04903b136067\") " pod="openshift-marketplace/redhat-operators-99bwv" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.582483 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63adda33-68e6-456d-beac-04903b136067-catalog-content\") pod \"redhat-operators-99bwv\" (UID: \"63adda33-68e6-456d-beac-04903b136067\") " pod="openshift-marketplace/redhat-operators-99bwv" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.582491 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63adda33-68e6-456d-beac-04903b136067-utilities\") pod \"redhat-operators-99bwv\" (UID: \"63adda33-68e6-456d-beac-04903b136067\") " pod="openshift-marketplace/redhat-operators-99bwv" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.595564 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.605714 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgctz\" (UniqueName: \"kubernetes.io/projected/63adda33-68e6-456d-beac-04903b136067-kube-api-access-jgctz\") pod \"redhat-operators-99bwv\" (UID: \"63adda33-68e6-456d-beac-04903b136067\") " pod="openshift-marketplace/redhat-operators-99bwv" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.653324 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.668507 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-99bwv" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.676830 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.888432 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-g9svq"] Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.889747 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-g9svq" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.895154 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.895578 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.899898 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-g9svq"] Nov 23 07:01:11 crc kubenswrapper[4559]: W1123 07:01:11.933264 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbaf469bd_b908_4271_8688_ee06c3b64e0b.slice/crio-0a80f4ecc8248d9b037e47e350adb54d45bebc4064bfbfb87d9a2f55830a8dde WatchSource:0}: Error finding container 0a80f4ecc8248d9b037e47e350adb54d45bebc4064bfbfb87d9a2f55830a8dde: Status 404 returned error can't find the container with id 0a80f4ecc8248d9b037e47e350adb54d45bebc4064bfbfb87d9a2f55830a8dde Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.934816 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-vfsc8"] Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.992598 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf4b8627-5670-4ab6-af0d-94b20fb48524-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-g9svq\" (UID: \"cf4b8627-5670-4ab6-af0d-94b20fb48524\") " pod="openstack/nova-cell1-conductor-db-sync-g9svq" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.994759 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf4b8627-5670-4ab6-af0d-94b20fb48524-scripts\") pod \"nova-cell1-conductor-db-sync-g9svq\" (UID: \"cf4b8627-5670-4ab6-af0d-94b20fb48524\") " pod="openstack/nova-cell1-conductor-db-sync-g9svq" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.997620 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4clz\" (UniqueName: \"kubernetes.io/projected/cf4b8627-5670-4ab6-af0d-94b20fb48524-kube-api-access-t4clz\") pod \"nova-cell1-conductor-db-sync-g9svq\" (UID: \"cf4b8627-5670-4ab6-af0d-94b20fb48524\") " pod="openstack/nova-cell1-conductor-db-sync-g9svq" Nov 23 07:01:11 crc kubenswrapper[4559]: I1123 07:01:11.997780 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf4b8627-5670-4ab6-af0d-94b20fb48524-config-data\") pod \"nova-cell1-conductor-db-sync-g9svq\" (UID: \"cf4b8627-5670-4ab6-af0d-94b20fb48524\") " pod="openstack/nova-cell1-conductor-db-sync-g9svq" Nov 23 07:01:12 crc kubenswrapper[4559]: I1123 07:01:12.099787 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf4b8627-5670-4ab6-af0d-94b20fb48524-scripts\") pod \"nova-cell1-conductor-db-sync-g9svq\" (UID: \"cf4b8627-5670-4ab6-af0d-94b20fb48524\") " pod="openstack/nova-cell1-conductor-db-sync-g9svq" Nov 23 07:01:12 crc kubenswrapper[4559]: I1123 07:01:12.099833 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4clz\" (UniqueName: \"kubernetes.io/projected/cf4b8627-5670-4ab6-af0d-94b20fb48524-kube-api-access-t4clz\") pod \"nova-cell1-conductor-db-sync-g9svq\" (UID: \"cf4b8627-5670-4ab6-af0d-94b20fb48524\") " pod="openstack/nova-cell1-conductor-db-sync-g9svq" Nov 23 07:01:12 crc kubenswrapper[4559]: I1123 07:01:12.099878 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf4b8627-5670-4ab6-af0d-94b20fb48524-config-data\") pod \"nova-cell1-conductor-db-sync-g9svq\" (UID: \"cf4b8627-5670-4ab6-af0d-94b20fb48524\") " pod="openstack/nova-cell1-conductor-db-sync-g9svq" Nov 23 07:01:12 crc kubenswrapper[4559]: I1123 07:01:12.099923 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf4b8627-5670-4ab6-af0d-94b20fb48524-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-g9svq\" (UID: \"cf4b8627-5670-4ab6-af0d-94b20fb48524\") " pod="openstack/nova-cell1-conductor-db-sync-g9svq" Nov 23 07:01:12 crc kubenswrapper[4559]: I1123 07:01:12.103615 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf4b8627-5670-4ab6-af0d-94b20fb48524-scripts\") pod \"nova-cell1-conductor-db-sync-g9svq\" (UID: \"cf4b8627-5670-4ab6-af0d-94b20fb48524\") " pod="openstack/nova-cell1-conductor-db-sync-g9svq" Nov 23 07:01:12 crc kubenswrapper[4559]: I1123 07:01:12.105428 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf4b8627-5670-4ab6-af0d-94b20fb48524-config-data\") pod \"nova-cell1-conductor-db-sync-g9svq\" (UID: \"cf4b8627-5670-4ab6-af0d-94b20fb48524\") " pod="openstack/nova-cell1-conductor-db-sync-g9svq" Nov 23 07:01:12 crc kubenswrapper[4559]: I1123 07:01:12.108056 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf4b8627-5670-4ab6-af0d-94b20fb48524-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-g9svq\" (UID: \"cf4b8627-5670-4ab6-af0d-94b20fb48524\") " pod="openstack/nova-cell1-conductor-db-sync-g9svq" Nov 23 07:01:12 crc kubenswrapper[4559]: I1123 07:01:12.119379 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4clz\" (UniqueName: \"kubernetes.io/projected/cf4b8627-5670-4ab6-af0d-94b20fb48524-kube-api-access-t4clz\") pod \"nova-cell1-conductor-db-sync-g9svq\" (UID: \"cf4b8627-5670-4ab6-af0d-94b20fb48524\") " pod="openstack/nova-cell1-conductor-db-sync-g9svq" Nov 23 07:01:12 crc kubenswrapper[4559]: I1123 07:01:12.146347 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:01:12 crc kubenswrapper[4559]: I1123 07:01:12.217589 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-g9svq" Nov 23 07:01:12 crc kubenswrapper[4559]: I1123 07:01:12.224029 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vfsc8" event={"ID":"baf469bd-b908-4271-8688-ee06c3b64e0b","Type":"ContainerStarted","Data":"0a80f4ecc8248d9b037e47e350adb54d45bebc4064bfbfb87d9a2f55830a8dde"} Nov 23 07:01:12 crc kubenswrapper[4559]: I1123 07:01:12.225046 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2522249b-6744-4b19-9ffd-d5dbad83a1bd","Type":"ContainerStarted","Data":"9c99d1a043cbf4416064f9518b9a4d1d2f84b633cc4d0cb9959b500422c105b5"} Nov 23 07:01:12 crc kubenswrapper[4559]: I1123 07:01:12.235609 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:01:12 crc kubenswrapper[4559]: I1123 07:01:12.366511 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 07:01:12 crc kubenswrapper[4559]: W1123 07:01:12.366824 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb7a9144_0988_44cc_9596_0cdf01b5abb3.slice/crio-e2ac0b4dc977aef988123030922249a9b4cbd6101644ad736ac41e2a268cb2ed WatchSource:0}: Error finding container e2ac0b4dc977aef988123030922249a9b4cbd6101644ad736ac41e2a268cb2ed: Status 404 returned error can't find the container with id e2ac0b4dc977aef988123030922249a9b4cbd6101644ad736ac41e2a268cb2ed Nov 23 07:01:12 crc kubenswrapper[4559]: W1123 07:01:12.459951 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod369ddeed_5e8f_47c2_9ea1_0f8f58ff74d5.slice/crio-54880d633a5c28b2087cd295aceed0ddc2aed986ccc64a2440448296a16fbd3c WatchSource:0}: Error finding container 54880d633a5c28b2087cd295aceed0ddc2aed986ccc64a2440448296a16fbd3c: Status 404 returned error can't find the container with id 54880d633a5c28b2087cd295aceed0ddc2aed986ccc64a2440448296a16fbd3c Nov 23 07:01:12 crc kubenswrapper[4559]: I1123 07:01:12.466766 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:01:12 crc kubenswrapper[4559]: I1123 07:01:12.489063 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-99bwv"] Nov 23 07:01:12 crc kubenswrapper[4559]: I1123 07:01:12.510080 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77cbccc799-6bxp4"] Nov 23 07:01:12 crc kubenswrapper[4559]: I1123 07:01:12.716037 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-g9svq"] Nov 23 07:01:12 crc kubenswrapper[4559]: W1123 07:01:12.743528 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf4b8627_5670_4ab6_af0d_94b20fb48524.slice/crio-154953fc25af0786fd0a2c8a4b62059a401f3b12b6fe6a3264c007a1ac68a608 WatchSource:0}: Error finding container 154953fc25af0786fd0a2c8a4b62059a401f3b12b6fe6a3264c007a1ac68a608: Status 404 returned error can't find the container with id 154953fc25af0786fd0a2c8a4b62059a401f3b12b6fe6a3264c007a1ac68a608 Nov 23 07:01:13 crc kubenswrapper[4559]: I1123 07:01:13.238402 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vfsc8" event={"ID":"baf469bd-b908-4271-8688-ee06c3b64e0b","Type":"ContainerStarted","Data":"ee299c98c771228f28a972b00f9944391ef80d73a6ae4a0822f84f31bcb2b1ed"} Nov 23 07:01:13 crc kubenswrapper[4559]: I1123 07:01:13.240172 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a05da27-e06e-4f41-91ae-ad450a604acb","Type":"ContainerStarted","Data":"e25ef91773e072a95cfeec45bc9a1f4254621971dd5ca4f9f619f445243cd89a"} Nov 23 07:01:13 crc kubenswrapper[4559]: I1123 07:01:13.242131 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bb7a9144-0988-44cc-9596-0cdf01b5abb3","Type":"ContainerStarted","Data":"e2ac0b4dc977aef988123030922249a9b4cbd6101644ad736ac41e2a268cb2ed"} Nov 23 07:01:13 crc kubenswrapper[4559]: I1123 07:01:13.245541 4559 generic.go:334] "Generic (PLEG): container finished" podID="9e626278-ea64-4839-8cf2-852c7d336dbb" containerID="558e702dfe65ebde1c55086eb1be211da8e0357ffe2640812361dc238e9ce254" exitCode=0 Nov 23 07:01:13 crc kubenswrapper[4559]: I1123 07:01:13.245605 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" event={"ID":"9e626278-ea64-4839-8cf2-852c7d336dbb","Type":"ContainerDied","Data":"558e702dfe65ebde1c55086eb1be211da8e0357ffe2640812361dc238e9ce254"} Nov 23 07:01:13 crc kubenswrapper[4559]: I1123 07:01:13.245631 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" event={"ID":"9e626278-ea64-4839-8cf2-852c7d336dbb","Type":"ContainerStarted","Data":"e0fbc0d60bd28f91402efdd653cf42c3c74f5270cb28b278880cc937e4fdf216"} Nov 23 07:01:13 crc kubenswrapper[4559]: I1123 07:01:13.249248 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-g9svq" event={"ID":"cf4b8627-5670-4ab6-af0d-94b20fb48524","Type":"ContainerStarted","Data":"eba4f1b416cd79fd936c7ba6b1670232b97f78c120cfd3d6633857972e6fe662"} Nov 23 07:01:13 crc kubenswrapper[4559]: I1123 07:01:13.249291 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-g9svq" event={"ID":"cf4b8627-5670-4ab6-af0d-94b20fb48524","Type":"ContainerStarted","Data":"154953fc25af0786fd0a2c8a4b62059a401f3b12b6fe6a3264c007a1ac68a608"} Nov 23 07:01:13 crc kubenswrapper[4559]: I1123 07:01:13.252506 4559 generic.go:334] "Generic (PLEG): container finished" podID="63adda33-68e6-456d-beac-04903b136067" containerID="8d0c7c6c63cdc10376edbec504e38196ad23ef4d45e38ff36e5688eb8e55b3ef" exitCode=0 Nov 23 07:01:13 crc kubenswrapper[4559]: I1123 07:01:13.252607 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-99bwv" event={"ID":"63adda33-68e6-456d-beac-04903b136067","Type":"ContainerDied","Data":"8d0c7c6c63cdc10376edbec504e38196ad23ef4d45e38ff36e5688eb8e55b3ef"} Nov 23 07:01:13 crc kubenswrapper[4559]: I1123 07:01:13.252680 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-99bwv" event={"ID":"63adda33-68e6-456d-beac-04903b136067","Type":"ContainerStarted","Data":"1a024a75fb8fcdd87f2ddc5b536131955d9d4e2bc289ea795108fa56e6cba5b6"} Nov 23 07:01:13 crc kubenswrapper[4559]: I1123 07:01:13.257605 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5","Type":"ContainerStarted","Data":"54880d633a5c28b2087cd295aceed0ddc2aed986ccc64a2440448296a16fbd3c"} Nov 23 07:01:13 crc kubenswrapper[4559]: I1123 07:01:13.259719 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-vfsc8" podStartSLOduration=3.25970769 podStartE2EDuration="3.25970769s" podCreationTimestamp="2025-11-23 07:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:01:13.249879878 +0000 UTC m=+975.271865492" watchObservedRunningTime="2025-11-23 07:01:13.25970769 +0000 UTC m=+975.281693304" Nov 23 07:01:13 crc kubenswrapper[4559]: I1123 07:01:13.280433 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-g9svq" podStartSLOduration=2.280416065 podStartE2EDuration="2.280416065s" podCreationTimestamp="2025-11-23 07:01:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:01:13.276022523 +0000 UTC m=+975.298008137" watchObservedRunningTime="2025-11-23 07:01:13.280416065 +0000 UTC m=+975.302401679" Nov 23 07:01:13 crc kubenswrapper[4559]: I1123 07:01:13.895815 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-n29p5"] Nov 23 07:01:13 crc kubenswrapper[4559]: I1123 07:01:13.909304 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n29p5" Nov 23 07:01:13 crc kubenswrapper[4559]: I1123 07:01:13.915481 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n29p5"] Nov 23 07:01:14 crc kubenswrapper[4559]: I1123 07:01:14.056218 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b00137d1-5df1-4e95-b066-827c4d2d53f8-catalog-content\") pod \"redhat-marketplace-n29p5\" (UID: \"b00137d1-5df1-4e95-b066-827c4d2d53f8\") " pod="openshift-marketplace/redhat-marketplace-n29p5" Nov 23 07:01:14 crc kubenswrapper[4559]: I1123 07:01:14.056388 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8qj7\" (UniqueName: \"kubernetes.io/projected/b00137d1-5df1-4e95-b066-827c4d2d53f8-kube-api-access-x8qj7\") pod \"redhat-marketplace-n29p5\" (UID: \"b00137d1-5df1-4e95-b066-827c4d2d53f8\") " pod="openshift-marketplace/redhat-marketplace-n29p5" Nov 23 07:01:14 crc kubenswrapper[4559]: I1123 07:01:14.056505 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b00137d1-5df1-4e95-b066-827c4d2d53f8-utilities\") pod \"redhat-marketplace-n29p5\" (UID: \"b00137d1-5df1-4e95-b066-827c4d2d53f8\") " pod="openshift-marketplace/redhat-marketplace-n29p5" Nov 23 07:01:14 crc kubenswrapper[4559]: I1123 07:01:14.159684 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b00137d1-5df1-4e95-b066-827c4d2d53f8-catalog-content\") pod \"redhat-marketplace-n29p5\" (UID: \"b00137d1-5df1-4e95-b066-827c4d2d53f8\") " pod="openshift-marketplace/redhat-marketplace-n29p5" Nov 23 07:01:14 crc kubenswrapper[4559]: I1123 07:01:14.159864 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8qj7\" (UniqueName: \"kubernetes.io/projected/b00137d1-5df1-4e95-b066-827c4d2d53f8-kube-api-access-x8qj7\") pod \"redhat-marketplace-n29p5\" (UID: \"b00137d1-5df1-4e95-b066-827c4d2d53f8\") " pod="openshift-marketplace/redhat-marketplace-n29p5" Nov 23 07:01:14 crc kubenswrapper[4559]: I1123 07:01:14.160042 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b00137d1-5df1-4e95-b066-827c4d2d53f8-utilities\") pod \"redhat-marketplace-n29p5\" (UID: \"b00137d1-5df1-4e95-b066-827c4d2d53f8\") " pod="openshift-marketplace/redhat-marketplace-n29p5" Nov 23 07:01:14 crc kubenswrapper[4559]: I1123 07:01:14.160515 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b00137d1-5df1-4e95-b066-827c4d2d53f8-catalog-content\") pod \"redhat-marketplace-n29p5\" (UID: \"b00137d1-5df1-4e95-b066-827c4d2d53f8\") " pod="openshift-marketplace/redhat-marketplace-n29p5" Nov 23 07:01:14 crc kubenswrapper[4559]: I1123 07:01:14.160573 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b00137d1-5df1-4e95-b066-827c4d2d53f8-utilities\") pod \"redhat-marketplace-n29p5\" (UID: \"b00137d1-5df1-4e95-b066-827c4d2d53f8\") " pod="openshift-marketplace/redhat-marketplace-n29p5" Nov 23 07:01:14 crc kubenswrapper[4559]: I1123 07:01:14.177903 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8qj7\" (UniqueName: \"kubernetes.io/projected/b00137d1-5df1-4e95-b066-827c4d2d53f8-kube-api-access-x8qj7\") pod \"redhat-marketplace-n29p5\" (UID: \"b00137d1-5df1-4e95-b066-827c4d2d53f8\") " pod="openshift-marketplace/redhat-marketplace-n29p5" Nov 23 07:01:14 crc kubenswrapper[4559]: I1123 07:01:14.262784 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n29p5" Nov 23 07:01:14 crc kubenswrapper[4559]: I1123 07:01:14.294606 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" event={"ID":"9e626278-ea64-4839-8cf2-852c7d336dbb","Type":"ContainerStarted","Data":"0deb60e939ea52ab20cf717534bd8464a2f2a8a8bde2111d095b2e77ce4b3cef"} Nov 23 07:01:14 crc kubenswrapper[4559]: I1123 07:01:14.294671 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" Nov 23 07:01:14 crc kubenswrapper[4559]: I1123 07:01:14.303289 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" podStartSLOduration=3.303273392 podStartE2EDuration="3.303273392s" podCreationTimestamp="2025-11-23 07:01:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:01:14.290832355 +0000 UTC m=+976.312817969" watchObservedRunningTime="2025-11-23 07:01:14.303273392 +0000 UTC m=+976.325259006" Nov 23 07:01:14 crc kubenswrapper[4559]: I1123 07:01:14.823624 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:01:14 crc kubenswrapper[4559]: I1123 07:01:14.834444 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 07:01:15 crc kubenswrapper[4559]: I1123 07:01:15.961448 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n29p5"] Nov 23 07:01:15 crc kubenswrapper[4559]: W1123 07:01:15.974991 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb00137d1_5df1_4e95_b066_827c4d2d53f8.slice/crio-405ae95273f331364a365dae793c720e1487e0a75de39534f940f909556d8ff7 WatchSource:0}: Error finding container 405ae95273f331364a365dae793c720e1487e0a75de39534f940f909556d8ff7: Status 404 returned error can't find the container with id 405ae95273f331364a365dae793c720e1487e0a75de39534f940f909556d8ff7 Nov 23 07:01:16 crc kubenswrapper[4559]: I1123 07:01:16.299080 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2522249b-6744-4b19-9ffd-d5dbad83a1bd","Type":"ContainerStarted","Data":"29142fa5f466a4fb22f0eab35c51f38525546c86f64092ff7e45dd685de3de3c"} Nov 23 07:01:16 crc kubenswrapper[4559]: I1123 07:01:16.301760 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-99bwv" event={"ID":"63adda33-68e6-456d-beac-04903b136067","Type":"ContainerStarted","Data":"bb5a9af84f00a0714495fb3402d6db3ed49864b36c7ea8f94ebfa038cbfb4fd2"} Nov 23 07:01:16 crc kubenswrapper[4559]: I1123 07:01:16.303154 4559 generic.go:334] "Generic (PLEG): container finished" podID="b00137d1-5df1-4e95-b066-827c4d2d53f8" containerID="d0ac0e16d4b8d3d3d2549819a456510f82fe5c4e2386c11d93e9431c732b82d7" exitCode=0 Nov 23 07:01:16 crc kubenswrapper[4559]: I1123 07:01:16.303207 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n29p5" event={"ID":"b00137d1-5df1-4e95-b066-827c4d2d53f8","Type":"ContainerDied","Data":"d0ac0e16d4b8d3d3d2549819a456510f82fe5c4e2386c11d93e9431c732b82d7"} Nov 23 07:01:16 crc kubenswrapper[4559]: I1123 07:01:16.303230 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n29p5" event={"ID":"b00137d1-5df1-4e95-b066-827c4d2d53f8","Type":"ContainerStarted","Data":"405ae95273f331364a365dae793c720e1487e0a75de39534f940f909556d8ff7"} Nov 23 07:01:16 crc kubenswrapper[4559]: I1123 07:01:16.304811 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5","Type":"ContainerStarted","Data":"c7d4b07cad1fdf517969aa5a8d5be54f2ba1878818411fd045b619b1b070e070"} Nov 23 07:01:16 crc kubenswrapper[4559]: I1123 07:01:16.304838 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5","Type":"ContainerStarted","Data":"4fe9ed5ce8ee873ef512afac3f47443d11fe31e55461e0d00f4e81bda788410e"} Nov 23 07:01:16 crc kubenswrapper[4559]: I1123 07:01:16.305025 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5" containerName="nova-metadata-log" containerID="cri-o://4fe9ed5ce8ee873ef512afac3f47443d11fe31e55461e0d00f4e81bda788410e" gracePeriod=30 Nov 23 07:01:16 crc kubenswrapper[4559]: I1123 07:01:16.305176 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5" containerName="nova-metadata-metadata" containerID="cri-o://c7d4b07cad1fdf517969aa5a8d5be54f2ba1878818411fd045b619b1b070e070" gracePeriod=30 Nov 23 07:01:16 crc kubenswrapper[4559]: I1123 07:01:16.311043 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a05da27-e06e-4f41-91ae-ad450a604acb","Type":"ContainerStarted","Data":"a1f386750a634bbeb64ad8ca451354cd825f3704999256eb110abb7eb6767023"} Nov 23 07:01:16 crc kubenswrapper[4559]: I1123 07:01:16.311138 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a05da27-e06e-4f41-91ae-ad450a604acb","Type":"ContainerStarted","Data":"219395f86bd287a5128cd7acaadcb2245bd1f872e19bc08b4210d42bc13ea36e"} Nov 23 07:01:16 crc kubenswrapper[4559]: I1123 07:01:16.320561 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bb7a9144-0988-44cc-9596-0cdf01b5abb3","Type":"ContainerStarted","Data":"37aff7304a6fa691f7155c46e4294d8257e6eaa652671e24f44cb8c31b40f0fe"} Nov 23 07:01:16 crc kubenswrapper[4559]: I1123 07:01:16.320696 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="bb7a9144-0988-44cc-9596-0cdf01b5abb3" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://37aff7304a6fa691f7155c46e4294d8257e6eaa652671e24f44cb8c31b40f0fe" gracePeriod=30 Nov 23 07:01:16 crc kubenswrapper[4559]: I1123 07:01:16.321361 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.040820196 podStartE2EDuration="5.321350151s" podCreationTimestamp="2025-11-23 07:01:11 +0000 UTC" firstStartedPulling="2025-11-23 07:01:12.1962077 +0000 UTC m=+974.218193313" lastFinishedPulling="2025-11-23 07:01:15.476737654 +0000 UTC m=+977.498723268" observedRunningTime="2025-11-23 07:01:16.31574535 +0000 UTC m=+978.337730964" watchObservedRunningTime="2025-11-23 07:01:16.321350151 +0000 UTC m=+978.343335765" Nov 23 07:01:16 crc kubenswrapper[4559]: I1123 07:01:16.377725 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.367340327 podStartE2EDuration="5.377700814s" podCreationTimestamp="2025-11-23 07:01:11 +0000 UTC" firstStartedPulling="2025-11-23 07:01:12.466216535 +0000 UTC m=+974.488202149" lastFinishedPulling="2025-11-23 07:01:15.476577021 +0000 UTC m=+977.498562636" observedRunningTime="2025-11-23 07:01:16.352574812 +0000 UTC m=+978.374560425" watchObservedRunningTime="2025-11-23 07:01:16.377700814 +0000 UTC m=+978.399686428" Nov 23 07:01:16 crc kubenswrapper[4559]: I1123 07:01:16.397391 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.172489607 podStartE2EDuration="6.397372179s" podCreationTimestamp="2025-11-23 07:01:10 +0000 UTC" firstStartedPulling="2025-11-23 07:01:12.25372081 +0000 UTC m=+974.275706425" lastFinishedPulling="2025-11-23 07:01:15.478603382 +0000 UTC m=+977.500588997" observedRunningTime="2025-11-23 07:01:16.386372633 +0000 UTC m=+978.408358247" watchObservedRunningTime="2025-11-23 07:01:16.397372179 +0000 UTC m=+978.419357793" Nov 23 07:01:16 crc kubenswrapper[4559]: I1123 07:01:16.409113 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.326330362 podStartE2EDuration="5.409101347s" podCreationTimestamp="2025-11-23 07:01:11 +0000 UTC" firstStartedPulling="2025-11-23 07:01:12.371143248 +0000 UTC m=+974.393128861" lastFinishedPulling="2025-11-23 07:01:15.453914232 +0000 UTC m=+977.475899846" observedRunningTime="2025-11-23 07:01:16.396684686 +0000 UTC m=+978.418670300" watchObservedRunningTime="2025-11-23 07:01:16.409101347 +0000 UTC m=+978.431086961" Nov 23 07:01:16 crc kubenswrapper[4559]: I1123 07:01:16.545039 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 23 07:01:16 crc kubenswrapper[4559]: I1123 07:01:16.654282 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:16 crc kubenswrapper[4559]: I1123 07:01:16.677456 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 23 07:01:16 crc kubenswrapper[4559]: I1123 07:01:16.677503 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 23 07:01:16 crc kubenswrapper[4559]: I1123 07:01:16.867522 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.049374 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5-logs\") pod \"369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5\" (UID: \"369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5\") " Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.049528 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5-combined-ca-bundle\") pod \"369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5\" (UID: \"369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5\") " Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.049726 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5-config-data\") pod \"369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5\" (UID: \"369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5\") " Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.049774 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfl89\" (UniqueName: \"kubernetes.io/projected/369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5-kube-api-access-rfl89\") pod \"369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5\" (UID: \"369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5\") " Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.049903 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5-logs" (OuterVolumeSpecName: "logs") pod "369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5" (UID: "369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.050480 4559 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.054992 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5-kube-api-access-rfl89" (OuterVolumeSpecName: "kube-api-access-rfl89") pod "369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5" (UID: "369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5"). InnerVolumeSpecName "kube-api-access-rfl89". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.091983 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5-config-data" (OuterVolumeSpecName: "config-data") pod "369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5" (UID: "369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.092211 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5" (UID: "369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.152252 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.152282 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.152293 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfl89\" (UniqueName: \"kubernetes.io/projected/369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5-kube-api-access-rfl89\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.332577 4559 generic.go:334] "Generic (PLEG): container finished" podID="cf4b8627-5670-4ab6-af0d-94b20fb48524" containerID="eba4f1b416cd79fd936c7ba6b1670232b97f78c120cfd3d6633857972e6fe662" exitCode=0 Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.332666 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-g9svq" event={"ID":"cf4b8627-5670-4ab6-af0d-94b20fb48524","Type":"ContainerDied","Data":"eba4f1b416cd79fd936c7ba6b1670232b97f78c120cfd3d6633857972e6fe662"} Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.335451 4559 generic.go:334] "Generic (PLEG): container finished" podID="63adda33-68e6-456d-beac-04903b136067" containerID="bb5a9af84f00a0714495fb3402d6db3ed49864b36c7ea8f94ebfa038cbfb4fd2" exitCode=0 Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.335603 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-99bwv" event={"ID":"63adda33-68e6-456d-beac-04903b136067","Type":"ContainerDied","Data":"bb5a9af84f00a0714495fb3402d6db3ed49864b36c7ea8f94ebfa038cbfb4fd2"} Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.339806 4559 generic.go:334] "Generic (PLEG): container finished" podID="b00137d1-5df1-4e95-b066-827c4d2d53f8" containerID="b731d59a4e4db9698acd3df6e767a006b3057c8bc2bd86a16fa6cbaecdbc5797" exitCode=0 Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.339864 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n29p5" event={"ID":"b00137d1-5df1-4e95-b066-827c4d2d53f8","Type":"ContainerDied","Data":"b731d59a4e4db9698acd3df6e767a006b3057c8bc2bd86a16fa6cbaecdbc5797"} Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.348725 4559 generic.go:334] "Generic (PLEG): container finished" podID="369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5" containerID="c7d4b07cad1fdf517969aa5a8d5be54f2ba1878818411fd045b619b1b070e070" exitCode=0 Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.348759 4559 generic.go:334] "Generic (PLEG): container finished" podID="369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5" containerID="4fe9ed5ce8ee873ef512afac3f47443d11fe31e55461e0d00f4e81bda788410e" exitCode=143 Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.349263 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.349321 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5","Type":"ContainerDied","Data":"c7d4b07cad1fdf517969aa5a8d5be54f2ba1878818411fd045b619b1b070e070"} Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.349346 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5","Type":"ContainerDied","Data":"4fe9ed5ce8ee873ef512afac3f47443d11fe31e55461e0d00f4e81bda788410e"} Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.349358 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5","Type":"ContainerDied","Data":"54880d633a5c28b2087cd295aceed0ddc2aed986ccc64a2440448296a16fbd3c"} Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.349374 4559 scope.go:117] "RemoveContainer" containerID="c7d4b07cad1fdf517969aa5a8d5be54f2ba1878818411fd045b619b1b070e070" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.379009 4559 scope.go:117] "RemoveContainer" containerID="4fe9ed5ce8ee873ef512afac3f47443d11fe31e55461e0d00f4e81bda788410e" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.416607 4559 scope.go:117] "RemoveContainer" containerID="c7d4b07cad1fdf517969aa5a8d5be54f2ba1878818411fd045b619b1b070e070" Nov 23 07:01:17 crc kubenswrapper[4559]: E1123 07:01:17.417122 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7d4b07cad1fdf517969aa5a8d5be54f2ba1878818411fd045b619b1b070e070\": container with ID starting with c7d4b07cad1fdf517969aa5a8d5be54f2ba1878818411fd045b619b1b070e070 not found: ID does not exist" containerID="c7d4b07cad1fdf517969aa5a8d5be54f2ba1878818411fd045b619b1b070e070" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.417170 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7d4b07cad1fdf517969aa5a8d5be54f2ba1878818411fd045b619b1b070e070"} err="failed to get container status \"c7d4b07cad1fdf517969aa5a8d5be54f2ba1878818411fd045b619b1b070e070\": rpc error: code = NotFound desc = could not find container \"c7d4b07cad1fdf517969aa5a8d5be54f2ba1878818411fd045b619b1b070e070\": container with ID starting with c7d4b07cad1fdf517969aa5a8d5be54f2ba1878818411fd045b619b1b070e070 not found: ID does not exist" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.417201 4559 scope.go:117] "RemoveContainer" containerID="4fe9ed5ce8ee873ef512afac3f47443d11fe31e55461e0d00f4e81bda788410e" Nov 23 07:01:17 crc kubenswrapper[4559]: E1123 07:01:17.417543 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fe9ed5ce8ee873ef512afac3f47443d11fe31e55461e0d00f4e81bda788410e\": container with ID starting with 4fe9ed5ce8ee873ef512afac3f47443d11fe31e55461e0d00f4e81bda788410e not found: ID does not exist" containerID="4fe9ed5ce8ee873ef512afac3f47443d11fe31e55461e0d00f4e81bda788410e" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.417608 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fe9ed5ce8ee873ef512afac3f47443d11fe31e55461e0d00f4e81bda788410e"} err="failed to get container status \"4fe9ed5ce8ee873ef512afac3f47443d11fe31e55461e0d00f4e81bda788410e\": rpc error: code = NotFound desc = could not find container \"4fe9ed5ce8ee873ef512afac3f47443d11fe31e55461e0d00f4e81bda788410e\": container with ID starting with 4fe9ed5ce8ee873ef512afac3f47443d11fe31e55461e0d00f4e81bda788410e not found: ID does not exist" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.417669 4559 scope.go:117] "RemoveContainer" containerID="c7d4b07cad1fdf517969aa5a8d5be54f2ba1878818411fd045b619b1b070e070" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.422996 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7d4b07cad1fdf517969aa5a8d5be54f2ba1878818411fd045b619b1b070e070"} err="failed to get container status \"c7d4b07cad1fdf517969aa5a8d5be54f2ba1878818411fd045b619b1b070e070\": rpc error: code = NotFound desc = could not find container \"c7d4b07cad1fdf517969aa5a8d5be54f2ba1878818411fd045b619b1b070e070\": container with ID starting with c7d4b07cad1fdf517969aa5a8d5be54f2ba1878818411fd045b619b1b070e070 not found: ID does not exist" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.423062 4559 scope.go:117] "RemoveContainer" containerID="4fe9ed5ce8ee873ef512afac3f47443d11fe31e55461e0d00f4e81bda788410e" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.423486 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fe9ed5ce8ee873ef512afac3f47443d11fe31e55461e0d00f4e81bda788410e"} err="failed to get container status \"4fe9ed5ce8ee873ef512afac3f47443d11fe31e55461e0d00f4e81bda788410e\": rpc error: code = NotFound desc = could not find container \"4fe9ed5ce8ee873ef512afac3f47443d11fe31e55461e0d00f4e81bda788410e\": container with ID starting with 4fe9ed5ce8ee873ef512afac3f47443d11fe31e55461e0d00f4e81bda788410e not found: ID does not exist" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.437703 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.472211 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.477539 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:01:17 crc kubenswrapper[4559]: E1123 07:01:17.477984 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5" containerName="nova-metadata-metadata" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.478003 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5" containerName="nova-metadata-metadata" Nov 23 07:01:17 crc kubenswrapper[4559]: E1123 07:01:17.478032 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5" containerName="nova-metadata-log" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.478047 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5" containerName="nova-metadata-log" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.478268 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5" containerName="nova-metadata-metadata" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.478288 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5" containerName="nova-metadata-log" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.479297 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.482724 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.483035 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.484359 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.563107 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45dfc36d-e083-4a20-bf1e-551cb08cdb77-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"45dfc36d-e083-4a20-bf1e-551cb08cdb77\") " pod="openstack/nova-metadata-0" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.563175 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/45dfc36d-e083-4a20-bf1e-551cb08cdb77-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"45dfc36d-e083-4a20-bf1e-551cb08cdb77\") " pod="openstack/nova-metadata-0" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.563263 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjsbn\" (UniqueName: \"kubernetes.io/projected/45dfc36d-e083-4a20-bf1e-551cb08cdb77-kube-api-access-bjsbn\") pod \"nova-metadata-0\" (UID: \"45dfc36d-e083-4a20-bf1e-551cb08cdb77\") " pod="openstack/nova-metadata-0" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.563333 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45dfc36d-e083-4a20-bf1e-551cb08cdb77-logs\") pod \"nova-metadata-0\" (UID: \"45dfc36d-e083-4a20-bf1e-551cb08cdb77\") " pod="openstack/nova-metadata-0" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.563481 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45dfc36d-e083-4a20-bf1e-551cb08cdb77-config-data\") pod \"nova-metadata-0\" (UID: \"45dfc36d-e083-4a20-bf1e-551cb08cdb77\") " pod="openstack/nova-metadata-0" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.665177 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/45dfc36d-e083-4a20-bf1e-551cb08cdb77-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"45dfc36d-e083-4a20-bf1e-551cb08cdb77\") " pod="openstack/nova-metadata-0" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.665307 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjsbn\" (UniqueName: \"kubernetes.io/projected/45dfc36d-e083-4a20-bf1e-551cb08cdb77-kube-api-access-bjsbn\") pod \"nova-metadata-0\" (UID: \"45dfc36d-e083-4a20-bf1e-551cb08cdb77\") " pod="openstack/nova-metadata-0" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.665396 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45dfc36d-e083-4a20-bf1e-551cb08cdb77-logs\") pod \"nova-metadata-0\" (UID: \"45dfc36d-e083-4a20-bf1e-551cb08cdb77\") " pod="openstack/nova-metadata-0" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.665427 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45dfc36d-e083-4a20-bf1e-551cb08cdb77-config-data\") pod \"nova-metadata-0\" (UID: \"45dfc36d-e083-4a20-bf1e-551cb08cdb77\") " pod="openstack/nova-metadata-0" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.665466 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45dfc36d-e083-4a20-bf1e-551cb08cdb77-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"45dfc36d-e083-4a20-bf1e-551cb08cdb77\") " pod="openstack/nova-metadata-0" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.665910 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45dfc36d-e083-4a20-bf1e-551cb08cdb77-logs\") pod \"nova-metadata-0\" (UID: \"45dfc36d-e083-4a20-bf1e-551cb08cdb77\") " pod="openstack/nova-metadata-0" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.669101 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45dfc36d-e083-4a20-bf1e-551cb08cdb77-config-data\") pod \"nova-metadata-0\" (UID: \"45dfc36d-e083-4a20-bf1e-551cb08cdb77\") " pod="openstack/nova-metadata-0" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.674879 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/45dfc36d-e083-4a20-bf1e-551cb08cdb77-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"45dfc36d-e083-4a20-bf1e-551cb08cdb77\") " pod="openstack/nova-metadata-0" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.678229 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45dfc36d-e083-4a20-bf1e-551cb08cdb77-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"45dfc36d-e083-4a20-bf1e-551cb08cdb77\") " pod="openstack/nova-metadata-0" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.692154 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjsbn\" (UniqueName: \"kubernetes.io/projected/45dfc36d-e083-4a20-bf1e-551cb08cdb77-kube-api-access-bjsbn\") pod \"nova-metadata-0\" (UID: \"45dfc36d-e083-4a20-bf1e-551cb08cdb77\") " pod="openstack/nova-metadata-0" Nov 23 07:01:17 crc kubenswrapper[4559]: I1123 07:01:17.808551 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:01:18 crc kubenswrapper[4559]: W1123 07:01:18.239215 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45dfc36d_e083_4a20_bf1e_551cb08cdb77.slice/crio-1bd44fd6c26d089efaa2c71eac392b6e9130a160ca8064954117926653c47c9c WatchSource:0}: Error finding container 1bd44fd6c26d089efaa2c71eac392b6e9130a160ca8064954117926653c47c9c: Status 404 returned error can't find the container with id 1bd44fd6c26d089efaa2c71eac392b6e9130a160ca8064954117926653c47c9c Nov 23 07:01:18 crc kubenswrapper[4559]: I1123 07:01:18.239226 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:01:18 crc kubenswrapper[4559]: I1123 07:01:18.284330 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5" path="/var/lib/kubelet/pods/369ddeed-5e8f-47c2-9ea1-0f8f58ff74d5/volumes" Nov 23 07:01:18 crc kubenswrapper[4559]: I1123 07:01:18.365376 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"45dfc36d-e083-4a20-bf1e-551cb08cdb77","Type":"ContainerStarted","Data":"1bd44fd6c26d089efaa2c71eac392b6e9130a160ca8064954117926653c47c9c"} Nov 23 07:01:18 crc kubenswrapper[4559]: I1123 07:01:18.368411 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-99bwv" event={"ID":"63adda33-68e6-456d-beac-04903b136067","Type":"ContainerStarted","Data":"bb49ec902c3bb819a91dd78740a3390349874432ece305c9d9017e7dd67bcf65"} Nov 23 07:01:18 crc kubenswrapper[4559]: I1123 07:01:18.370789 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n29p5" event={"ID":"b00137d1-5df1-4e95-b066-827c4d2d53f8","Type":"ContainerStarted","Data":"56272bdfc95aa42872687268934b43ac5627b3a728fa2921f39a8e17eefc381f"} Nov 23 07:01:18 crc kubenswrapper[4559]: I1123 07:01:18.392597 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-99bwv" podStartSLOduration=2.786079767 podStartE2EDuration="7.39258053s" podCreationTimestamp="2025-11-23 07:01:11 +0000 UTC" firstStartedPulling="2025-11-23 07:01:13.256603091 +0000 UTC m=+975.278588706" lastFinishedPulling="2025-11-23 07:01:17.863103855 +0000 UTC m=+979.885089469" observedRunningTime="2025-11-23 07:01:18.383325896 +0000 UTC m=+980.405311510" watchObservedRunningTime="2025-11-23 07:01:18.39258053 +0000 UTC m=+980.414566144" Nov 23 07:01:18 crc kubenswrapper[4559]: I1123 07:01:18.422365 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-n29p5" podStartSLOduration=3.867514177 podStartE2EDuration="5.422341308s" podCreationTimestamp="2025-11-23 07:01:13 +0000 UTC" firstStartedPulling="2025-11-23 07:01:16.308435783 +0000 UTC m=+978.330421397" lastFinishedPulling="2025-11-23 07:01:17.863262914 +0000 UTC m=+979.885248528" observedRunningTime="2025-11-23 07:01:18.398853507 +0000 UTC m=+980.420839122" watchObservedRunningTime="2025-11-23 07:01:18.422341308 +0000 UTC m=+980.444326923" Nov 23 07:01:18 crc kubenswrapper[4559]: I1123 07:01:18.708624 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-g9svq" Nov 23 07:01:18 crc kubenswrapper[4559]: I1123 07:01:18.892584 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4clz\" (UniqueName: \"kubernetes.io/projected/cf4b8627-5670-4ab6-af0d-94b20fb48524-kube-api-access-t4clz\") pod \"cf4b8627-5670-4ab6-af0d-94b20fb48524\" (UID: \"cf4b8627-5670-4ab6-af0d-94b20fb48524\") " Nov 23 07:01:18 crc kubenswrapper[4559]: I1123 07:01:18.892695 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf4b8627-5670-4ab6-af0d-94b20fb48524-config-data\") pod \"cf4b8627-5670-4ab6-af0d-94b20fb48524\" (UID: \"cf4b8627-5670-4ab6-af0d-94b20fb48524\") " Nov 23 07:01:18 crc kubenswrapper[4559]: I1123 07:01:18.892849 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf4b8627-5670-4ab6-af0d-94b20fb48524-scripts\") pod \"cf4b8627-5670-4ab6-af0d-94b20fb48524\" (UID: \"cf4b8627-5670-4ab6-af0d-94b20fb48524\") " Nov 23 07:01:18 crc kubenswrapper[4559]: I1123 07:01:18.893475 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf4b8627-5670-4ab6-af0d-94b20fb48524-combined-ca-bundle\") pod \"cf4b8627-5670-4ab6-af0d-94b20fb48524\" (UID: \"cf4b8627-5670-4ab6-af0d-94b20fb48524\") " Nov 23 07:01:18 crc kubenswrapper[4559]: I1123 07:01:18.898254 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf4b8627-5670-4ab6-af0d-94b20fb48524-scripts" (OuterVolumeSpecName: "scripts") pod "cf4b8627-5670-4ab6-af0d-94b20fb48524" (UID: "cf4b8627-5670-4ab6-af0d-94b20fb48524"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:18 crc kubenswrapper[4559]: I1123 07:01:18.898343 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf4b8627-5670-4ab6-af0d-94b20fb48524-kube-api-access-t4clz" (OuterVolumeSpecName: "kube-api-access-t4clz") pod "cf4b8627-5670-4ab6-af0d-94b20fb48524" (UID: "cf4b8627-5670-4ab6-af0d-94b20fb48524"). InnerVolumeSpecName "kube-api-access-t4clz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:01:18 crc kubenswrapper[4559]: I1123 07:01:18.939213 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf4b8627-5670-4ab6-af0d-94b20fb48524-config-data" (OuterVolumeSpecName: "config-data") pod "cf4b8627-5670-4ab6-af0d-94b20fb48524" (UID: "cf4b8627-5670-4ab6-af0d-94b20fb48524"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:18 crc kubenswrapper[4559]: I1123 07:01:18.948188 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf4b8627-5670-4ab6-af0d-94b20fb48524-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cf4b8627-5670-4ab6-af0d-94b20fb48524" (UID: "cf4b8627-5670-4ab6-af0d-94b20fb48524"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:18 crc kubenswrapper[4559]: I1123 07:01:18.996308 4559 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf4b8627-5670-4ab6-af0d-94b20fb48524-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:18 crc kubenswrapper[4559]: I1123 07:01:18.996341 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf4b8627-5670-4ab6-af0d-94b20fb48524-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:18 crc kubenswrapper[4559]: I1123 07:01:18.996354 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4clz\" (UniqueName: \"kubernetes.io/projected/cf4b8627-5670-4ab6-af0d-94b20fb48524-kube-api-access-t4clz\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:18 crc kubenswrapper[4559]: I1123 07:01:18.996363 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf4b8627-5670-4ab6-af0d-94b20fb48524-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:19 crc kubenswrapper[4559]: I1123 07:01:19.380379 4559 generic.go:334] "Generic (PLEG): container finished" podID="baf469bd-b908-4271-8688-ee06c3b64e0b" containerID="ee299c98c771228f28a972b00f9944391ef80d73a6ae4a0822f84f31bcb2b1ed" exitCode=0 Nov 23 07:01:19 crc kubenswrapper[4559]: I1123 07:01:19.380457 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vfsc8" event={"ID":"baf469bd-b908-4271-8688-ee06c3b64e0b","Type":"ContainerDied","Data":"ee299c98c771228f28a972b00f9944391ef80d73a6ae4a0822f84f31bcb2b1ed"} Nov 23 07:01:19 crc kubenswrapper[4559]: I1123 07:01:19.383261 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"45dfc36d-e083-4a20-bf1e-551cb08cdb77","Type":"ContainerStarted","Data":"7bc52602b529d29b818f4745f8349ff1dacac7052047f4501f3bb1133fe7582f"} Nov 23 07:01:19 crc kubenswrapper[4559]: I1123 07:01:19.383294 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"45dfc36d-e083-4a20-bf1e-551cb08cdb77","Type":"ContainerStarted","Data":"98cac2efc1626ab1f57d0abaf6efb9be3ae5fc9929e313e0a6c56cb807e2222e"} Nov 23 07:01:19 crc kubenswrapper[4559]: I1123 07:01:19.385167 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-g9svq" event={"ID":"cf4b8627-5670-4ab6-af0d-94b20fb48524","Type":"ContainerDied","Data":"154953fc25af0786fd0a2c8a4b62059a401f3b12b6fe6a3264c007a1ac68a608"} Nov 23 07:01:19 crc kubenswrapper[4559]: I1123 07:01:19.385205 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="154953fc25af0786fd0a2c8a4b62059a401f3b12b6fe6a3264c007a1ac68a608" Nov 23 07:01:19 crc kubenswrapper[4559]: I1123 07:01:19.385264 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-g9svq" Nov 23 07:01:19 crc kubenswrapper[4559]: I1123 07:01:19.418130 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.4181059449999998 podStartE2EDuration="2.418105945s" podCreationTimestamp="2025-11-23 07:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:01:19.414767837 +0000 UTC m=+981.436753450" watchObservedRunningTime="2025-11-23 07:01:19.418105945 +0000 UTC m=+981.440091549" Nov 23 07:01:19 crc kubenswrapper[4559]: I1123 07:01:19.419833 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 23 07:01:19 crc kubenswrapper[4559]: E1123 07:01:19.420275 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf4b8627-5670-4ab6-af0d-94b20fb48524" containerName="nova-cell1-conductor-db-sync" Nov 23 07:01:19 crc kubenswrapper[4559]: I1123 07:01:19.420294 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf4b8627-5670-4ab6-af0d-94b20fb48524" containerName="nova-cell1-conductor-db-sync" Nov 23 07:01:19 crc kubenswrapper[4559]: I1123 07:01:19.420464 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf4b8627-5670-4ab6-af0d-94b20fb48524" containerName="nova-cell1-conductor-db-sync" Nov 23 07:01:19 crc kubenswrapper[4559]: I1123 07:01:19.421132 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 23 07:01:19 crc kubenswrapper[4559]: I1123 07:01:19.423002 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 23 07:01:19 crc kubenswrapper[4559]: I1123 07:01:19.435982 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 23 07:01:19 crc kubenswrapper[4559]: I1123 07:01:19.609807 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56c491cd-4182-4477-a9de-d7e91d68e07f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"56c491cd-4182-4477-a9de-d7e91d68e07f\") " pod="openstack/nova-cell1-conductor-0" Nov 23 07:01:19 crc kubenswrapper[4559]: I1123 07:01:19.610152 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdrff\" (UniqueName: \"kubernetes.io/projected/56c491cd-4182-4477-a9de-d7e91d68e07f-kube-api-access-bdrff\") pod \"nova-cell1-conductor-0\" (UID: \"56c491cd-4182-4477-a9de-d7e91d68e07f\") " pod="openstack/nova-cell1-conductor-0" Nov 23 07:01:19 crc kubenswrapper[4559]: I1123 07:01:19.610388 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56c491cd-4182-4477-a9de-d7e91d68e07f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"56c491cd-4182-4477-a9de-d7e91d68e07f\") " pod="openstack/nova-cell1-conductor-0" Nov 23 07:01:19 crc kubenswrapper[4559]: I1123 07:01:19.712414 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56c491cd-4182-4477-a9de-d7e91d68e07f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"56c491cd-4182-4477-a9de-d7e91d68e07f\") " pod="openstack/nova-cell1-conductor-0" Nov 23 07:01:19 crc kubenswrapper[4559]: I1123 07:01:19.712481 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56c491cd-4182-4477-a9de-d7e91d68e07f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"56c491cd-4182-4477-a9de-d7e91d68e07f\") " pod="openstack/nova-cell1-conductor-0" Nov 23 07:01:19 crc kubenswrapper[4559]: I1123 07:01:19.712548 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdrff\" (UniqueName: \"kubernetes.io/projected/56c491cd-4182-4477-a9de-d7e91d68e07f-kube-api-access-bdrff\") pod \"nova-cell1-conductor-0\" (UID: \"56c491cd-4182-4477-a9de-d7e91d68e07f\") " pod="openstack/nova-cell1-conductor-0" Nov 23 07:01:19 crc kubenswrapper[4559]: I1123 07:01:19.718365 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56c491cd-4182-4477-a9de-d7e91d68e07f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"56c491cd-4182-4477-a9de-d7e91d68e07f\") " pod="openstack/nova-cell1-conductor-0" Nov 23 07:01:19 crc kubenswrapper[4559]: I1123 07:01:19.718365 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56c491cd-4182-4477-a9de-d7e91d68e07f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"56c491cd-4182-4477-a9de-d7e91d68e07f\") " pod="openstack/nova-cell1-conductor-0" Nov 23 07:01:19 crc kubenswrapper[4559]: I1123 07:01:19.727601 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdrff\" (UniqueName: \"kubernetes.io/projected/56c491cd-4182-4477-a9de-d7e91d68e07f-kube-api-access-bdrff\") pod \"nova-cell1-conductor-0\" (UID: \"56c491cd-4182-4477-a9de-d7e91d68e07f\") " pod="openstack/nova-cell1-conductor-0" Nov 23 07:01:19 crc kubenswrapper[4559]: I1123 07:01:19.739530 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 23 07:01:20 crc kubenswrapper[4559]: I1123 07:01:20.166708 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 23 07:01:20 crc kubenswrapper[4559]: I1123 07:01:20.394164 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"56c491cd-4182-4477-a9de-d7e91d68e07f","Type":"ContainerStarted","Data":"8e25951fc95058cda5c08cfc96422e2771eaf5d160dca0b3e4c26eb7e81a5bc9"} Nov 23 07:01:20 crc kubenswrapper[4559]: I1123 07:01:20.395174 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"56c491cd-4182-4477-a9de-d7e91d68e07f","Type":"ContainerStarted","Data":"14e2be1c8b4d833b679cd570240a4c1cbd7c004d2e7a53cc27138c68b58918f3"} Nov 23 07:01:20 crc kubenswrapper[4559]: I1123 07:01:20.395195 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 23 07:01:20 crc kubenswrapper[4559]: I1123 07:01:20.411360 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=1.411339761 podStartE2EDuration="1.411339761s" podCreationTimestamp="2025-11-23 07:01:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:01:20.408610569 +0000 UTC m=+982.430596193" watchObservedRunningTime="2025-11-23 07:01:20.411339761 +0000 UTC m=+982.433325375" Nov 23 07:01:20 crc kubenswrapper[4559]: I1123 07:01:20.752985 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vfsc8" Nov 23 07:01:20 crc kubenswrapper[4559]: I1123 07:01:20.843201 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baf469bd-b908-4271-8688-ee06c3b64e0b-scripts\") pod \"baf469bd-b908-4271-8688-ee06c3b64e0b\" (UID: \"baf469bd-b908-4271-8688-ee06c3b64e0b\") " Nov 23 07:01:20 crc kubenswrapper[4559]: I1123 07:01:20.843320 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfhlg\" (UniqueName: \"kubernetes.io/projected/baf469bd-b908-4271-8688-ee06c3b64e0b-kube-api-access-lfhlg\") pod \"baf469bd-b908-4271-8688-ee06c3b64e0b\" (UID: \"baf469bd-b908-4271-8688-ee06c3b64e0b\") " Nov 23 07:01:20 crc kubenswrapper[4559]: I1123 07:01:20.843450 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baf469bd-b908-4271-8688-ee06c3b64e0b-combined-ca-bundle\") pod \"baf469bd-b908-4271-8688-ee06c3b64e0b\" (UID: \"baf469bd-b908-4271-8688-ee06c3b64e0b\") " Nov 23 07:01:20 crc kubenswrapper[4559]: I1123 07:01:20.843538 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baf469bd-b908-4271-8688-ee06c3b64e0b-config-data\") pod \"baf469bd-b908-4271-8688-ee06c3b64e0b\" (UID: \"baf469bd-b908-4271-8688-ee06c3b64e0b\") " Nov 23 07:01:20 crc kubenswrapper[4559]: I1123 07:01:20.848913 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/baf469bd-b908-4271-8688-ee06c3b64e0b-kube-api-access-lfhlg" (OuterVolumeSpecName: "kube-api-access-lfhlg") pod "baf469bd-b908-4271-8688-ee06c3b64e0b" (UID: "baf469bd-b908-4271-8688-ee06c3b64e0b"). InnerVolumeSpecName "kube-api-access-lfhlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:01:20 crc kubenswrapper[4559]: I1123 07:01:20.850811 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baf469bd-b908-4271-8688-ee06c3b64e0b-scripts" (OuterVolumeSpecName: "scripts") pod "baf469bd-b908-4271-8688-ee06c3b64e0b" (UID: "baf469bd-b908-4271-8688-ee06c3b64e0b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:20 crc kubenswrapper[4559]: E1123 07:01:20.866417 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/baf469bd-b908-4271-8688-ee06c3b64e0b-config-data podName:baf469bd-b908-4271-8688-ee06c3b64e0b nodeName:}" failed. No retries permitted until 2025-11-23 07:01:21.366380149 +0000 UTC m=+983.388365764 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/baf469bd-b908-4271-8688-ee06c3b64e0b-config-data") pod "baf469bd-b908-4271-8688-ee06c3b64e0b" (UID: "baf469bd-b908-4271-8688-ee06c3b64e0b") : error deleting /var/lib/kubelet/pods/baf469bd-b908-4271-8688-ee06c3b64e0b/volume-subpaths: remove /var/lib/kubelet/pods/baf469bd-b908-4271-8688-ee06c3b64e0b/volume-subpaths: no such file or directory Nov 23 07:01:20 crc kubenswrapper[4559]: I1123 07:01:20.869908 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baf469bd-b908-4271-8688-ee06c3b64e0b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "baf469bd-b908-4271-8688-ee06c3b64e0b" (UID: "baf469bd-b908-4271-8688-ee06c3b64e0b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:20 crc kubenswrapper[4559]: I1123 07:01:20.946594 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfhlg\" (UniqueName: \"kubernetes.io/projected/baf469bd-b908-4271-8688-ee06c3b64e0b-kube-api-access-lfhlg\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:20 crc kubenswrapper[4559]: I1123 07:01:20.946628 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baf469bd-b908-4271-8688-ee06c3b64e0b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:20 crc kubenswrapper[4559]: I1123 07:01:20.946654 4559 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baf469bd-b908-4271-8688-ee06c3b64e0b-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:21 crc kubenswrapper[4559]: I1123 07:01:21.407103 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vfsc8" Nov 23 07:01:21 crc kubenswrapper[4559]: I1123 07:01:21.407677 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vfsc8" event={"ID":"baf469bd-b908-4271-8688-ee06c3b64e0b","Type":"ContainerDied","Data":"0a80f4ecc8248d9b037e47e350adb54d45bebc4064bfbfb87d9a2f55830a8dde"} Nov 23 07:01:21 crc kubenswrapper[4559]: I1123 07:01:21.407713 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a80f4ecc8248d9b037e47e350adb54d45bebc4064bfbfb87d9a2f55830a8dde" Nov 23 07:01:21 crc kubenswrapper[4559]: I1123 07:01:21.457281 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baf469bd-b908-4271-8688-ee06c3b64e0b-config-data\") pod \"baf469bd-b908-4271-8688-ee06c3b64e0b\" (UID: \"baf469bd-b908-4271-8688-ee06c3b64e0b\") " Nov 23 07:01:21 crc kubenswrapper[4559]: I1123 07:01:21.460760 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baf469bd-b908-4271-8688-ee06c3b64e0b-config-data" (OuterVolumeSpecName: "config-data") pod "baf469bd-b908-4271-8688-ee06c3b64e0b" (UID: "baf469bd-b908-4271-8688-ee06c3b64e0b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:21 crc kubenswrapper[4559]: I1123 07:01:21.544757 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 23 07:01:21 crc kubenswrapper[4559]: I1123 07:01:21.560626 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baf469bd-b908-4271-8688-ee06c3b64e0b-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:21 crc kubenswrapper[4559]: I1123 07:01:21.569839 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" Nov 23 07:01:21 crc kubenswrapper[4559]: I1123 07:01:21.603340 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:01:21 crc kubenswrapper[4559]: I1123 07:01:21.603584 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0a05da27-e06e-4f41-91ae-ad450a604acb" containerName="nova-api-log" containerID="cri-o://219395f86bd287a5128cd7acaadcb2245bd1f872e19bc08b4210d42bc13ea36e" gracePeriod=30 Nov 23 07:01:21 crc kubenswrapper[4559]: I1123 07:01:21.603805 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0a05da27-e06e-4f41-91ae-ad450a604acb" containerName="nova-api-api" containerID="cri-o://a1f386750a634bbeb64ad8ca451354cd825f3704999256eb110abb7eb6767023" gracePeriod=30 Nov 23 07:01:21 crc kubenswrapper[4559]: I1123 07:01:21.616772 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:01:21 crc kubenswrapper[4559]: I1123 07:01:21.623133 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 23 07:01:21 crc kubenswrapper[4559]: I1123 07:01:21.623760 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:01:21 crc kubenswrapper[4559]: I1123 07:01:21.623978 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="45dfc36d-e083-4a20-bf1e-551cb08cdb77" containerName="nova-metadata-log" containerID="cri-o://98cac2efc1626ab1f57d0abaf6efb9be3ae5fc9929e313e0a6c56cb807e2222e" gracePeriod=30 Nov 23 07:01:21 crc kubenswrapper[4559]: I1123 07:01:21.624015 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="45dfc36d-e083-4a20-bf1e-551cb08cdb77" containerName="nova-metadata-metadata" containerID="cri-o://7bc52602b529d29b818f4745f8349ff1dacac7052047f4501f3bb1133fe7582f" gracePeriod=30 Nov 23 07:01:21 crc kubenswrapper[4559]: I1123 07:01:21.657619 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64cc7f6975-rdztg"] Nov 23 07:01:21 crc kubenswrapper[4559]: I1123 07:01:21.658147 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" podUID="d6acba84-09d4-4657-8c25-b863919294a2" containerName="dnsmasq-dns" containerID="cri-o://b6e390e295765cffce68ec1cc51febc51204bd83a3fd7e2e7ea92c138ecbaa16" gracePeriod=10 Nov 23 07:01:21 crc kubenswrapper[4559]: I1123 07:01:21.670136 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-99bwv" Nov 23 07:01:21 crc kubenswrapper[4559]: I1123 07:01:21.670169 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-99bwv" Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.435184 4559 generic.go:334] "Generic (PLEG): container finished" podID="45dfc36d-e083-4a20-bf1e-551cb08cdb77" containerID="7bc52602b529d29b818f4745f8349ff1dacac7052047f4501f3bb1133fe7582f" exitCode=0 Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.435445 4559 generic.go:334] "Generic (PLEG): container finished" podID="45dfc36d-e083-4a20-bf1e-551cb08cdb77" containerID="98cac2efc1626ab1f57d0abaf6efb9be3ae5fc9929e313e0a6c56cb807e2222e" exitCode=143 Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.435347 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"45dfc36d-e083-4a20-bf1e-551cb08cdb77","Type":"ContainerDied","Data":"7bc52602b529d29b818f4745f8349ff1dacac7052047f4501f3bb1133fe7582f"} Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.435519 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"45dfc36d-e083-4a20-bf1e-551cb08cdb77","Type":"ContainerDied","Data":"98cac2efc1626ab1f57d0abaf6efb9be3ae5fc9929e313e0a6c56cb807e2222e"} Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.437544 4559 generic.go:334] "Generic (PLEG): container finished" podID="d6acba84-09d4-4657-8c25-b863919294a2" containerID="b6e390e295765cffce68ec1cc51febc51204bd83a3fd7e2e7ea92c138ecbaa16" exitCode=0 Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.437600 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" event={"ID":"d6acba84-09d4-4657-8c25-b863919294a2","Type":"ContainerDied","Data":"b6e390e295765cffce68ec1cc51febc51204bd83a3fd7e2e7ea92c138ecbaa16"} Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.439447 4559 generic.go:334] "Generic (PLEG): container finished" podID="0a05da27-e06e-4f41-91ae-ad450a604acb" containerID="a1f386750a634bbeb64ad8ca451354cd825f3704999256eb110abb7eb6767023" exitCode=0 Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.439469 4559 generic.go:334] "Generic (PLEG): container finished" podID="0a05da27-e06e-4f41-91ae-ad450a604acb" containerID="219395f86bd287a5128cd7acaadcb2245bd1f872e19bc08b4210d42bc13ea36e" exitCode=143 Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.439606 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="2522249b-6744-4b19-9ffd-d5dbad83a1bd" containerName="nova-scheduler-scheduler" containerID="cri-o://29142fa5f466a4fb22f0eab35c51f38525546c86f64092ff7e45dd685de3de3c" gracePeriod=30 Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.439875 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a05da27-e06e-4f41-91ae-ad450a604acb","Type":"ContainerDied","Data":"a1f386750a634bbeb64ad8ca451354cd825f3704999256eb110abb7eb6767023"} Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.439901 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a05da27-e06e-4f41-91ae-ad450a604acb","Type":"ContainerDied","Data":"219395f86bd287a5128cd7acaadcb2245bd1f872e19bc08b4210d42bc13ea36e"} Nov 23 07:01:22 crc kubenswrapper[4559]: E1123 07:01:22.442245 4559 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="29142fa5f466a4fb22f0eab35c51f38525546c86f64092ff7e45dd685de3de3c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 23 07:01:22 crc kubenswrapper[4559]: E1123 07:01:22.443314 4559 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="29142fa5f466a4fb22f0eab35c51f38525546c86f64092ff7e45dd685de3de3c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 23 07:01:22 crc kubenswrapper[4559]: E1123 07:01:22.446667 4559 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="29142fa5f466a4fb22f0eab35c51f38525546c86f64092ff7e45dd685de3de3c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 23 07:01:22 crc kubenswrapper[4559]: E1123 07:01:22.446717 4559 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="2522249b-6744-4b19-9ffd-d5dbad83a1bd" containerName="nova-scheduler-scheduler" Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.620196 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.683038 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-dns-svc\") pod \"d6acba84-09d4-4657-8c25-b863919294a2\" (UID: \"d6acba84-09d4-4657-8c25-b863919294a2\") " Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.683084 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5dgf\" (UniqueName: \"kubernetes.io/projected/d6acba84-09d4-4657-8c25-b863919294a2-kube-api-access-d5dgf\") pod \"d6acba84-09d4-4657-8c25-b863919294a2\" (UID: \"d6acba84-09d4-4657-8c25-b863919294a2\") " Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.683136 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-config\") pod \"d6acba84-09d4-4657-8c25-b863919294a2\" (UID: \"d6acba84-09d4-4657-8c25-b863919294a2\") " Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.683219 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-dns-swift-storage-0\") pod \"d6acba84-09d4-4657-8c25-b863919294a2\" (UID: \"d6acba84-09d4-4657-8c25-b863919294a2\") " Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.683268 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-ovsdbserver-sb\") pod \"d6acba84-09d4-4657-8c25-b863919294a2\" (UID: \"d6acba84-09d4-4657-8c25-b863919294a2\") " Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.683304 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-ovsdbserver-nb\") pod \"d6acba84-09d4-4657-8c25-b863919294a2\" (UID: \"d6acba84-09d4-4657-8c25-b863919294a2\") " Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.697071 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6acba84-09d4-4657-8c25-b863919294a2-kube-api-access-d5dgf" (OuterVolumeSpecName: "kube-api-access-d5dgf") pod "d6acba84-09d4-4657-8c25-b863919294a2" (UID: "d6acba84-09d4-4657-8c25-b863919294a2"). InnerVolumeSpecName "kube-api-access-d5dgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.736364 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-99bwv" podUID="63adda33-68e6-456d-beac-04903b136067" containerName="registry-server" probeResult="failure" output=< Nov 23 07:01:22 crc kubenswrapper[4559]: timeout: failed to connect service ":50051" within 1s Nov 23 07:01:22 crc kubenswrapper[4559]: > Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.740347 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d6acba84-09d4-4657-8c25-b863919294a2" (UID: "d6acba84-09d4-4657-8c25-b863919294a2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.741768 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d6acba84-09d4-4657-8c25-b863919294a2" (UID: "d6acba84-09d4-4657-8c25-b863919294a2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.757632 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-config" (OuterVolumeSpecName: "config") pod "d6acba84-09d4-4657-8c25-b863919294a2" (UID: "d6acba84-09d4-4657-8c25-b863919294a2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.757931 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d6acba84-09d4-4657-8c25-b863919294a2" (UID: "d6acba84-09d4-4657-8c25-b863919294a2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.760633 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d6acba84-09d4-4657-8c25-b863919294a2" (UID: "d6acba84-09d4-4657-8c25-b863919294a2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.786092 4559 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.786123 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5dgf\" (UniqueName: \"kubernetes.io/projected/d6acba84-09d4-4657-8c25-b863919294a2-kube-api-access-d5dgf\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.786136 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.786145 4559 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.786154 4559 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.786162 4559 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d6acba84-09d4-4657-8c25-b863919294a2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.808735 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.808804 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.950256 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.958341 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.988610 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/45dfc36d-e083-4a20-bf1e-551cb08cdb77-nova-metadata-tls-certs\") pod \"45dfc36d-e083-4a20-bf1e-551cb08cdb77\" (UID: \"45dfc36d-e083-4a20-bf1e-551cb08cdb77\") " Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.988657 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a05da27-e06e-4f41-91ae-ad450a604acb-config-data\") pod \"0a05da27-e06e-4f41-91ae-ad450a604acb\" (UID: \"0a05da27-e06e-4f41-91ae-ad450a604acb\") " Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.988678 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45dfc36d-e083-4a20-bf1e-551cb08cdb77-config-data\") pod \"45dfc36d-e083-4a20-bf1e-551cb08cdb77\" (UID: \"45dfc36d-e083-4a20-bf1e-551cb08cdb77\") " Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.988712 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a05da27-e06e-4f41-91ae-ad450a604acb-combined-ca-bundle\") pod \"0a05da27-e06e-4f41-91ae-ad450a604acb\" (UID: \"0a05da27-e06e-4f41-91ae-ad450a604acb\") " Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.988788 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d92qt\" (UniqueName: \"kubernetes.io/projected/0a05da27-e06e-4f41-91ae-ad450a604acb-kube-api-access-d92qt\") pod \"0a05da27-e06e-4f41-91ae-ad450a604acb\" (UID: \"0a05da27-e06e-4f41-91ae-ad450a604acb\") " Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.988849 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a05da27-e06e-4f41-91ae-ad450a604acb-logs\") pod \"0a05da27-e06e-4f41-91ae-ad450a604acb\" (UID: \"0a05da27-e06e-4f41-91ae-ad450a604acb\") " Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.988868 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45dfc36d-e083-4a20-bf1e-551cb08cdb77-logs\") pod \"45dfc36d-e083-4a20-bf1e-551cb08cdb77\" (UID: \"45dfc36d-e083-4a20-bf1e-551cb08cdb77\") " Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.988893 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45dfc36d-e083-4a20-bf1e-551cb08cdb77-combined-ca-bundle\") pod \"45dfc36d-e083-4a20-bf1e-551cb08cdb77\" (UID: \"45dfc36d-e083-4a20-bf1e-551cb08cdb77\") " Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.988912 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjsbn\" (UniqueName: \"kubernetes.io/projected/45dfc36d-e083-4a20-bf1e-551cb08cdb77-kube-api-access-bjsbn\") pod \"45dfc36d-e083-4a20-bf1e-551cb08cdb77\" (UID: \"45dfc36d-e083-4a20-bf1e-551cb08cdb77\") " Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.989567 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a05da27-e06e-4f41-91ae-ad450a604acb-logs" (OuterVolumeSpecName: "logs") pod "0a05da27-e06e-4f41-91ae-ad450a604acb" (UID: "0a05da27-e06e-4f41-91ae-ad450a604acb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.991962 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45dfc36d-e083-4a20-bf1e-551cb08cdb77-logs" (OuterVolumeSpecName: "logs") pod "45dfc36d-e083-4a20-bf1e-551cb08cdb77" (UID: "45dfc36d-e083-4a20-bf1e-551cb08cdb77"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:01:22 crc kubenswrapper[4559]: I1123 07:01:22.993808 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45dfc36d-e083-4a20-bf1e-551cb08cdb77-kube-api-access-bjsbn" (OuterVolumeSpecName: "kube-api-access-bjsbn") pod "45dfc36d-e083-4a20-bf1e-551cb08cdb77" (UID: "45dfc36d-e083-4a20-bf1e-551cb08cdb77"). InnerVolumeSpecName "kube-api-access-bjsbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.009918 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a05da27-e06e-4f41-91ae-ad450a604acb-kube-api-access-d92qt" (OuterVolumeSpecName: "kube-api-access-d92qt") pod "0a05da27-e06e-4f41-91ae-ad450a604acb" (UID: "0a05da27-e06e-4f41-91ae-ad450a604acb"). InnerVolumeSpecName "kube-api-access-d92qt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.014175 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a05da27-e06e-4f41-91ae-ad450a604acb-config-data" (OuterVolumeSpecName: "config-data") pod "0a05da27-e06e-4f41-91ae-ad450a604acb" (UID: "0a05da27-e06e-4f41-91ae-ad450a604acb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.018286 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45dfc36d-e083-4a20-bf1e-551cb08cdb77-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "45dfc36d-e083-4a20-bf1e-551cb08cdb77" (UID: "45dfc36d-e083-4a20-bf1e-551cb08cdb77"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.024783 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45dfc36d-e083-4a20-bf1e-551cb08cdb77-config-data" (OuterVolumeSpecName: "config-data") pod "45dfc36d-e083-4a20-bf1e-551cb08cdb77" (UID: "45dfc36d-e083-4a20-bf1e-551cb08cdb77"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.026171 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a05da27-e06e-4f41-91ae-ad450a604acb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a05da27-e06e-4f41-91ae-ad450a604acb" (UID: "0a05da27-e06e-4f41-91ae-ad450a604acb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.049223 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45dfc36d-e083-4a20-bf1e-551cb08cdb77-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "45dfc36d-e083-4a20-bf1e-551cb08cdb77" (UID: "45dfc36d-e083-4a20-bf1e-551cb08cdb77"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.091028 4559 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/45dfc36d-e083-4a20-bf1e-551cb08cdb77-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.091377 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a05da27-e06e-4f41-91ae-ad450a604acb-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.091388 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45dfc36d-e083-4a20-bf1e-551cb08cdb77-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.091397 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a05da27-e06e-4f41-91ae-ad450a604acb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.091406 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d92qt\" (UniqueName: \"kubernetes.io/projected/0a05da27-e06e-4f41-91ae-ad450a604acb-kube-api-access-d92qt\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.091415 4559 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a05da27-e06e-4f41-91ae-ad450a604acb-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.091422 4559 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45dfc36d-e083-4a20-bf1e-551cb08cdb77-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.091430 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45dfc36d-e083-4a20-bf1e-551cb08cdb77-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.091438 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjsbn\" (UniqueName: \"kubernetes.io/projected/45dfc36d-e083-4a20-bf1e-551cb08cdb77-kube-api-access-bjsbn\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.452783 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" event={"ID":"d6acba84-09d4-4657-8c25-b863919294a2","Type":"ContainerDied","Data":"6ef48fdd1d3bfc863106c8e1d3c7fbe097bc254d1d11a662310b29bac0e63730"} Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.452853 4559 scope.go:117] "RemoveContainer" containerID="b6e390e295765cffce68ec1cc51febc51204bd83a3fd7e2e7ea92c138ecbaa16" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.452868 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64cc7f6975-rdztg" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.460168 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a05da27-e06e-4f41-91ae-ad450a604acb","Type":"ContainerDied","Data":"e25ef91773e072a95cfeec45bc9a1f4254621971dd5ca4f9f619f445243cd89a"} Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.460283 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.466879 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"45dfc36d-e083-4a20-bf1e-551cb08cdb77","Type":"ContainerDied","Data":"1bd44fd6c26d089efaa2c71eac392b6e9130a160ca8064954117926653c47c9c"} Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.466987 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.531109 4559 scope.go:117] "RemoveContainer" containerID="401abf9d16ce3ec67e1fcbd50bb57d6e050b1ae263b46da20fbb84f857f9bb65" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.536516 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.565409 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.569929 4559 scope.go:117] "RemoveContainer" containerID="a1f386750a634bbeb64ad8ca451354cd825f3704999256eb110abb7eb6767023" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.586153 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64cc7f6975-rdztg"] Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.607356 4559 scope.go:117] "RemoveContainer" containerID="219395f86bd287a5128cd7acaadcb2245bd1f872e19bc08b4210d42bc13ea36e" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.607489 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:01:23 crc kubenswrapper[4559]: E1123 07:01:23.607936 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45dfc36d-e083-4a20-bf1e-551cb08cdb77" containerName="nova-metadata-metadata" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.607952 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="45dfc36d-e083-4a20-bf1e-551cb08cdb77" containerName="nova-metadata-metadata" Nov 23 07:01:23 crc kubenswrapper[4559]: E1123 07:01:23.607982 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a05da27-e06e-4f41-91ae-ad450a604acb" containerName="nova-api-api" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.607989 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a05da27-e06e-4f41-91ae-ad450a604acb" containerName="nova-api-api" Nov 23 07:01:23 crc kubenswrapper[4559]: E1123 07:01:23.608129 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baf469bd-b908-4271-8688-ee06c3b64e0b" containerName="nova-manage" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.608172 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="baf469bd-b908-4271-8688-ee06c3b64e0b" containerName="nova-manage" Nov 23 07:01:23 crc kubenswrapper[4559]: E1123 07:01:23.608198 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6acba84-09d4-4657-8c25-b863919294a2" containerName="dnsmasq-dns" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.608206 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6acba84-09d4-4657-8c25-b863919294a2" containerName="dnsmasq-dns" Nov 23 07:01:23 crc kubenswrapper[4559]: E1123 07:01:23.608269 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6acba84-09d4-4657-8c25-b863919294a2" containerName="init" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.608278 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6acba84-09d4-4657-8c25-b863919294a2" containerName="init" Nov 23 07:01:23 crc kubenswrapper[4559]: E1123 07:01:23.608292 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a05da27-e06e-4f41-91ae-ad450a604acb" containerName="nova-api-log" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.608299 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a05da27-e06e-4f41-91ae-ad450a604acb" containerName="nova-api-log" Nov 23 07:01:23 crc kubenswrapper[4559]: E1123 07:01:23.608308 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45dfc36d-e083-4a20-bf1e-551cb08cdb77" containerName="nova-metadata-log" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.608315 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="45dfc36d-e083-4a20-bf1e-551cb08cdb77" containerName="nova-metadata-log" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.608675 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="45dfc36d-e083-4a20-bf1e-551cb08cdb77" containerName="nova-metadata-metadata" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.608701 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a05da27-e06e-4f41-91ae-ad450a604acb" containerName="nova-api-log" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.608710 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a05da27-e06e-4f41-91ae-ad450a604acb" containerName="nova-api-api" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.608721 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6acba84-09d4-4657-8c25-b863919294a2" containerName="dnsmasq-dns" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.608731 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="baf469bd-b908-4271-8688-ee06c3b64e0b" containerName="nova-manage" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.608757 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="45dfc36d-e083-4a20-bf1e-551cb08cdb77" containerName="nova-metadata-log" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.609931 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.612304 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.612351 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.616716 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64cc7f6975-rdztg"] Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.623724 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.630899 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.631395 4559 scope.go:117] "RemoveContainer" containerID="7bc52602b529d29b818f4745f8349ff1dacac7052047f4501f3bb1133fe7582f" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.636545 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.641493 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.643175 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.649164 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.650449 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.652909 4559 scope.go:117] "RemoveContainer" containerID="98cac2efc1626ab1f57d0abaf6efb9be3ae5fc9929e313e0a6c56cb807e2222e" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.700989 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d0d4e86-3876-4666-b628-83cb6c08700b-config-data\") pod \"nova-metadata-0\" (UID: \"8d0d4e86-3876-4666-b628-83cb6c08700b\") " pod="openstack/nova-metadata-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.701033 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b0260a5-8273-4d39-a06f-b1d02ff9696e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7b0260a5-8273-4d39-a06f-b1d02ff9696e\") " pod="openstack/nova-api-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.701056 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lv65h\" (UniqueName: \"kubernetes.io/projected/7b0260a5-8273-4d39-a06f-b1d02ff9696e-kube-api-access-lv65h\") pod \"nova-api-0\" (UID: \"7b0260a5-8273-4d39-a06f-b1d02ff9696e\") " pod="openstack/nova-api-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.701348 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b0260a5-8273-4d39-a06f-b1d02ff9696e-logs\") pod \"nova-api-0\" (UID: \"7b0260a5-8273-4d39-a06f-b1d02ff9696e\") " pod="openstack/nova-api-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.701443 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d0d4e86-3876-4666-b628-83cb6c08700b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8d0d4e86-3876-4666-b628-83cb6c08700b\") " pod="openstack/nova-metadata-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.701508 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d0d4e86-3876-4666-b628-83cb6c08700b-logs\") pod \"nova-metadata-0\" (UID: \"8d0d4e86-3876-4666-b628-83cb6c08700b\") " pod="openstack/nova-metadata-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.701752 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b0260a5-8273-4d39-a06f-b1d02ff9696e-config-data\") pod \"nova-api-0\" (UID: \"7b0260a5-8273-4d39-a06f-b1d02ff9696e\") " pod="openstack/nova-api-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.701848 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7n5n\" (UniqueName: \"kubernetes.io/projected/8d0d4e86-3876-4666-b628-83cb6c08700b-kube-api-access-p7n5n\") pod \"nova-metadata-0\" (UID: \"8d0d4e86-3876-4666-b628-83cb6c08700b\") " pod="openstack/nova-metadata-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.701871 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d0d4e86-3876-4666-b628-83cb6c08700b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8d0d4e86-3876-4666-b628-83cb6c08700b\") " pod="openstack/nova-metadata-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.802467 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b0260a5-8273-4d39-a06f-b1d02ff9696e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7b0260a5-8273-4d39-a06f-b1d02ff9696e\") " pod="openstack/nova-api-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.802497 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lv65h\" (UniqueName: \"kubernetes.io/projected/7b0260a5-8273-4d39-a06f-b1d02ff9696e-kube-api-access-lv65h\") pod \"nova-api-0\" (UID: \"7b0260a5-8273-4d39-a06f-b1d02ff9696e\") " pod="openstack/nova-api-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.802582 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b0260a5-8273-4d39-a06f-b1d02ff9696e-logs\") pod \"nova-api-0\" (UID: \"7b0260a5-8273-4d39-a06f-b1d02ff9696e\") " pod="openstack/nova-api-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.802618 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d0d4e86-3876-4666-b628-83cb6c08700b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8d0d4e86-3876-4666-b628-83cb6c08700b\") " pod="openstack/nova-metadata-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.802657 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d0d4e86-3876-4666-b628-83cb6c08700b-logs\") pod \"nova-metadata-0\" (UID: \"8d0d4e86-3876-4666-b628-83cb6c08700b\") " pod="openstack/nova-metadata-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.802678 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b0260a5-8273-4d39-a06f-b1d02ff9696e-config-data\") pod \"nova-api-0\" (UID: \"7b0260a5-8273-4d39-a06f-b1d02ff9696e\") " pod="openstack/nova-api-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.802710 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7n5n\" (UniqueName: \"kubernetes.io/projected/8d0d4e86-3876-4666-b628-83cb6c08700b-kube-api-access-p7n5n\") pod \"nova-metadata-0\" (UID: \"8d0d4e86-3876-4666-b628-83cb6c08700b\") " pod="openstack/nova-metadata-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.802728 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d0d4e86-3876-4666-b628-83cb6c08700b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8d0d4e86-3876-4666-b628-83cb6c08700b\") " pod="openstack/nova-metadata-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.802761 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d0d4e86-3876-4666-b628-83cb6c08700b-config-data\") pod \"nova-metadata-0\" (UID: \"8d0d4e86-3876-4666-b628-83cb6c08700b\") " pod="openstack/nova-metadata-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.803164 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d0d4e86-3876-4666-b628-83cb6c08700b-logs\") pod \"nova-metadata-0\" (UID: \"8d0d4e86-3876-4666-b628-83cb6c08700b\") " pod="openstack/nova-metadata-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.803198 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b0260a5-8273-4d39-a06f-b1d02ff9696e-logs\") pod \"nova-api-0\" (UID: \"7b0260a5-8273-4d39-a06f-b1d02ff9696e\") " pod="openstack/nova-api-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.807026 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d0d4e86-3876-4666-b628-83cb6c08700b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8d0d4e86-3876-4666-b628-83cb6c08700b\") " pod="openstack/nova-metadata-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.808078 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d0d4e86-3876-4666-b628-83cb6c08700b-config-data\") pod \"nova-metadata-0\" (UID: \"8d0d4e86-3876-4666-b628-83cb6c08700b\") " pod="openstack/nova-metadata-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.808266 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b0260a5-8273-4d39-a06f-b1d02ff9696e-config-data\") pod \"nova-api-0\" (UID: \"7b0260a5-8273-4d39-a06f-b1d02ff9696e\") " pod="openstack/nova-api-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.808609 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d0d4e86-3876-4666-b628-83cb6c08700b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8d0d4e86-3876-4666-b628-83cb6c08700b\") " pod="openstack/nova-metadata-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.810009 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b0260a5-8273-4d39-a06f-b1d02ff9696e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7b0260a5-8273-4d39-a06f-b1d02ff9696e\") " pod="openstack/nova-api-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.817620 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7n5n\" (UniqueName: \"kubernetes.io/projected/8d0d4e86-3876-4666-b628-83cb6c08700b-kube-api-access-p7n5n\") pod \"nova-metadata-0\" (UID: \"8d0d4e86-3876-4666-b628-83cb6c08700b\") " pod="openstack/nova-metadata-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.818102 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lv65h\" (UniqueName: \"kubernetes.io/projected/7b0260a5-8273-4d39-a06f-b1d02ff9696e-kube-api-access-lv65h\") pod \"nova-api-0\" (UID: \"7b0260a5-8273-4d39-a06f-b1d02ff9696e\") " pod="openstack/nova-api-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.933187 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:01:23 crc kubenswrapper[4559]: I1123 07:01:23.961931 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:01:24 crc kubenswrapper[4559]: I1123 07:01:24.263588 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-n29p5" Nov 23 07:01:24 crc kubenswrapper[4559]: I1123 07:01:24.264071 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-n29p5" Nov 23 07:01:24 crc kubenswrapper[4559]: I1123 07:01:24.285703 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a05da27-e06e-4f41-91ae-ad450a604acb" path="/var/lib/kubelet/pods/0a05da27-e06e-4f41-91ae-ad450a604acb/volumes" Nov 23 07:01:24 crc kubenswrapper[4559]: I1123 07:01:24.286614 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45dfc36d-e083-4a20-bf1e-551cb08cdb77" path="/var/lib/kubelet/pods/45dfc36d-e083-4a20-bf1e-551cb08cdb77/volumes" Nov 23 07:01:24 crc kubenswrapper[4559]: I1123 07:01:24.287216 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6acba84-09d4-4657-8c25-b863919294a2" path="/var/lib/kubelet/pods/d6acba84-09d4-4657-8c25-b863919294a2/volumes" Nov 23 07:01:24 crc kubenswrapper[4559]: I1123 07:01:24.304224 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-n29p5" Nov 23 07:01:24 crc kubenswrapper[4559]: I1123 07:01:24.347308 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:01:24 crc kubenswrapper[4559]: W1123 07:01:24.348223 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d0d4e86_3876_4666_b628_83cb6c08700b.slice/crio-9882b4ab9934b0c3bf41b995a669faf80fc2133ae12977ecaa8cb978f65d6754 WatchSource:0}: Error finding container 9882b4ab9934b0c3bf41b995a669faf80fc2133ae12977ecaa8cb978f65d6754: Status 404 returned error can't find the container with id 9882b4ab9934b0c3bf41b995a669faf80fc2133ae12977ecaa8cb978f65d6754 Nov 23 07:01:24 crc kubenswrapper[4559]: I1123 07:01:24.405291 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:01:24 crc kubenswrapper[4559]: W1123 07:01:24.405402 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b0260a5_8273_4d39_a06f_b1d02ff9696e.slice/crio-d13a722723e93a451712aeff55e4d7d5aa41689c6a976caedb38131bdb4231ff WatchSource:0}: Error finding container d13a722723e93a451712aeff55e4d7d5aa41689c6a976caedb38131bdb4231ff: Status 404 returned error can't find the container with id d13a722723e93a451712aeff55e4d7d5aa41689c6a976caedb38131bdb4231ff Nov 23 07:01:24 crc kubenswrapper[4559]: I1123 07:01:24.477759 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8d0d4e86-3876-4666-b628-83cb6c08700b","Type":"ContainerStarted","Data":"022d9852ba23da5dec14c28273df4974709ac8978ff19f432268af796578b513"} Nov 23 07:01:24 crc kubenswrapper[4559]: I1123 07:01:24.477798 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8d0d4e86-3876-4666-b628-83cb6c08700b","Type":"ContainerStarted","Data":"9882b4ab9934b0c3bf41b995a669faf80fc2133ae12977ecaa8cb978f65d6754"} Nov 23 07:01:24 crc kubenswrapper[4559]: I1123 07:01:24.481625 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7b0260a5-8273-4d39-a06f-b1d02ff9696e","Type":"ContainerStarted","Data":"d13a722723e93a451712aeff55e4d7d5aa41689c6a976caedb38131bdb4231ff"} Nov 23 07:01:24 crc kubenswrapper[4559]: I1123 07:01:24.516311 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-n29p5" Nov 23 07:01:24 crc kubenswrapper[4559]: I1123 07:01:24.570054 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n29p5"] Nov 23 07:01:25 crc kubenswrapper[4559]: I1123 07:01:25.490633 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7b0260a5-8273-4d39-a06f-b1d02ff9696e","Type":"ContainerStarted","Data":"2a5677e783c8afdf414c7343add7bd352ebead848a4aff47acdacb3119347ebe"} Nov 23 07:01:25 crc kubenswrapper[4559]: I1123 07:01:25.491000 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7b0260a5-8273-4d39-a06f-b1d02ff9696e","Type":"ContainerStarted","Data":"75e3395816a677c9b178c87af51011cedca02a3b7d307f0a6cb1e864de5b6886"} Nov 23 07:01:25 crc kubenswrapper[4559]: I1123 07:01:25.492553 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8d0d4e86-3876-4666-b628-83cb6c08700b","Type":"ContainerStarted","Data":"9b5836b9d27c68a11951faf7967525bf1cd78e4a672b484c343e84c3363bac02"} Nov 23 07:01:25 crc kubenswrapper[4559]: I1123 07:01:25.509332 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.509321024 podStartE2EDuration="2.509321024s" podCreationTimestamp="2025-11-23 07:01:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:01:25.505243805 +0000 UTC m=+987.527229420" watchObservedRunningTime="2025-11-23 07:01:25.509321024 +0000 UTC m=+987.531306637" Nov 23 07:01:25 crc kubenswrapper[4559]: I1123 07:01:25.528467 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.528455147 podStartE2EDuration="2.528455147s" podCreationTimestamp="2025-11-23 07:01:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:01:25.520832321 +0000 UTC m=+987.542817935" watchObservedRunningTime="2025-11-23 07:01:25.528455147 +0000 UTC m=+987.550440752" Nov 23 07:01:26 crc kubenswrapper[4559]: I1123 07:01:26.504785 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-n29p5" podUID="b00137d1-5df1-4e95-b066-827c4d2d53f8" containerName="registry-server" containerID="cri-o://56272bdfc95aa42872687268934b43ac5627b3a728fa2921f39a8e17eefc381f" gracePeriod=2 Nov 23 07:01:26 crc kubenswrapper[4559]: E1123 07:01:26.546421 4559 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="29142fa5f466a4fb22f0eab35c51f38525546c86f64092ff7e45dd685de3de3c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 23 07:01:26 crc kubenswrapper[4559]: E1123 07:01:26.547946 4559 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="29142fa5f466a4fb22f0eab35c51f38525546c86f64092ff7e45dd685de3de3c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 23 07:01:26 crc kubenswrapper[4559]: E1123 07:01:26.549082 4559 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="29142fa5f466a4fb22f0eab35c51f38525546c86f64092ff7e45dd685de3de3c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 23 07:01:26 crc kubenswrapper[4559]: E1123 07:01:26.549144 4559 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="2522249b-6744-4b19-9ffd-d5dbad83a1bd" containerName="nova-scheduler-scheduler" Nov 23 07:01:26 crc kubenswrapper[4559]: I1123 07:01:26.916404 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n29p5" Nov 23 07:01:26 crc kubenswrapper[4559]: I1123 07:01:26.971907 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b00137d1-5df1-4e95-b066-827c4d2d53f8-catalog-content\") pod \"b00137d1-5df1-4e95-b066-827c4d2d53f8\" (UID: \"b00137d1-5df1-4e95-b066-827c4d2d53f8\") " Nov 23 07:01:26 crc kubenswrapper[4559]: I1123 07:01:26.972622 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b00137d1-5df1-4e95-b066-827c4d2d53f8-utilities\") pod \"b00137d1-5df1-4e95-b066-827c4d2d53f8\" (UID: \"b00137d1-5df1-4e95-b066-827c4d2d53f8\") " Nov 23 07:01:26 crc kubenswrapper[4559]: I1123 07:01:26.972749 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8qj7\" (UniqueName: \"kubernetes.io/projected/b00137d1-5df1-4e95-b066-827c4d2d53f8-kube-api-access-x8qj7\") pod \"b00137d1-5df1-4e95-b066-827c4d2d53f8\" (UID: \"b00137d1-5df1-4e95-b066-827c4d2d53f8\") " Nov 23 07:01:26 crc kubenswrapper[4559]: I1123 07:01:26.973563 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b00137d1-5df1-4e95-b066-827c4d2d53f8-utilities" (OuterVolumeSpecName: "utilities") pod "b00137d1-5df1-4e95-b066-827c4d2d53f8" (UID: "b00137d1-5df1-4e95-b066-827c4d2d53f8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:01:26 crc kubenswrapper[4559]: I1123 07:01:26.979320 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b00137d1-5df1-4e95-b066-827c4d2d53f8-kube-api-access-x8qj7" (OuterVolumeSpecName: "kube-api-access-x8qj7") pod "b00137d1-5df1-4e95-b066-827c4d2d53f8" (UID: "b00137d1-5df1-4e95-b066-827c4d2d53f8"). InnerVolumeSpecName "kube-api-access-x8qj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:01:26 crc kubenswrapper[4559]: I1123 07:01:26.988583 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b00137d1-5df1-4e95-b066-827c4d2d53f8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b00137d1-5df1-4e95-b066-827c4d2d53f8" (UID: "b00137d1-5df1-4e95-b066-827c4d2d53f8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.075593 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b00137d1-5df1-4e95-b066-827c4d2d53f8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.075744 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b00137d1-5df1-4e95-b066-827c4d2d53f8-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.075825 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8qj7\" (UniqueName: \"kubernetes.io/projected/b00137d1-5df1-4e95-b066-827c4d2d53f8-kube-api-access-x8qj7\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.293247 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.380177 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87pcm\" (UniqueName: \"kubernetes.io/projected/2522249b-6744-4b19-9ffd-d5dbad83a1bd-kube-api-access-87pcm\") pod \"2522249b-6744-4b19-9ffd-d5dbad83a1bd\" (UID: \"2522249b-6744-4b19-9ffd-d5dbad83a1bd\") " Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.384717 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2522249b-6744-4b19-9ffd-d5dbad83a1bd-kube-api-access-87pcm" (OuterVolumeSpecName: "kube-api-access-87pcm") pod "2522249b-6744-4b19-9ffd-d5dbad83a1bd" (UID: "2522249b-6744-4b19-9ffd-d5dbad83a1bd"). InnerVolumeSpecName "kube-api-access-87pcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.481536 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2522249b-6744-4b19-9ffd-d5dbad83a1bd-combined-ca-bundle\") pod \"2522249b-6744-4b19-9ffd-d5dbad83a1bd\" (UID: \"2522249b-6744-4b19-9ffd-d5dbad83a1bd\") " Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.481722 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2522249b-6744-4b19-9ffd-d5dbad83a1bd-config-data\") pod \"2522249b-6744-4b19-9ffd-d5dbad83a1bd\" (UID: \"2522249b-6744-4b19-9ffd-d5dbad83a1bd\") " Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.482969 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87pcm\" (UniqueName: \"kubernetes.io/projected/2522249b-6744-4b19-9ffd-d5dbad83a1bd-kube-api-access-87pcm\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.509291 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2522249b-6744-4b19-9ffd-d5dbad83a1bd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2522249b-6744-4b19-9ffd-d5dbad83a1bd" (UID: "2522249b-6744-4b19-9ffd-d5dbad83a1bd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.509306 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2522249b-6744-4b19-9ffd-d5dbad83a1bd-config-data" (OuterVolumeSpecName: "config-data") pod "2522249b-6744-4b19-9ffd-d5dbad83a1bd" (UID: "2522249b-6744-4b19-9ffd-d5dbad83a1bd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.513798 4559 generic.go:334] "Generic (PLEG): container finished" podID="2522249b-6744-4b19-9ffd-d5dbad83a1bd" containerID="29142fa5f466a4fb22f0eab35c51f38525546c86f64092ff7e45dd685de3de3c" exitCode=0 Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.513843 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2522249b-6744-4b19-9ffd-d5dbad83a1bd","Type":"ContainerDied","Data":"29142fa5f466a4fb22f0eab35c51f38525546c86f64092ff7e45dd685de3de3c"} Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.513891 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2522249b-6744-4b19-9ffd-d5dbad83a1bd","Type":"ContainerDied","Data":"9c99d1a043cbf4416064f9518b9a4d1d2f84b633cc4d0cb9959b500422c105b5"} Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.513915 4559 scope.go:117] "RemoveContainer" containerID="29142fa5f466a4fb22f0eab35c51f38525546c86f64092ff7e45dd685de3de3c" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.513955 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.518496 4559 generic.go:334] "Generic (PLEG): container finished" podID="b00137d1-5df1-4e95-b066-827c4d2d53f8" containerID="56272bdfc95aa42872687268934b43ac5627b3a728fa2921f39a8e17eefc381f" exitCode=0 Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.518811 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n29p5" event={"ID":"b00137d1-5df1-4e95-b066-827c4d2d53f8","Type":"ContainerDied","Data":"56272bdfc95aa42872687268934b43ac5627b3a728fa2921f39a8e17eefc381f"} Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.518831 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n29p5" event={"ID":"b00137d1-5df1-4e95-b066-827c4d2d53f8","Type":"ContainerDied","Data":"405ae95273f331364a365dae793c720e1487e0a75de39534f940f909556d8ff7"} Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.518893 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n29p5" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.538447 4559 scope.go:117] "RemoveContainer" containerID="29142fa5f466a4fb22f0eab35c51f38525546c86f64092ff7e45dd685de3de3c" Nov 23 07:01:27 crc kubenswrapper[4559]: E1123 07:01:27.543296 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29142fa5f466a4fb22f0eab35c51f38525546c86f64092ff7e45dd685de3de3c\": container with ID starting with 29142fa5f466a4fb22f0eab35c51f38525546c86f64092ff7e45dd685de3de3c not found: ID does not exist" containerID="29142fa5f466a4fb22f0eab35c51f38525546c86f64092ff7e45dd685de3de3c" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.543325 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29142fa5f466a4fb22f0eab35c51f38525546c86f64092ff7e45dd685de3de3c"} err="failed to get container status \"29142fa5f466a4fb22f0eab35c51f38525546c86f64092ff7e45dd685de3de3c\": rpc error: code = NotFound desc = could not find container \"29142fa5f466a4fb22f0eab35c51f38525546c86f64092ff7e45dd685de3de3c\": container with ID starting with 29142fa5f466a4fb22f0eab35c51f38525546c86f64092ff7e45dd685de3de3c not found: ID does not exist" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.543348 4559 scope.go:117] "RemoveContainer" containerID="56272bdfc95aa42872687268934b43ac5627b3a728fa2921f39a8e17eefc381f" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.550168 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.561319 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.565157 4559 scope.go:117] "RemoveContainer" containerID="b731d59a4e4db9698acd3df6e767a006b3057c8bc2bd86a16fa6cbaecdbc5797" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.575858 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n29p5"] Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.584136 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-n29p5"] Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.585042 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2522249b-6744-4b19-9ffd-d5dbad83a1bd-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.585061 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2522249b-6744-4b19-9ffd-d5dbad83a1bd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.589104 4559 scope.go:117] "RemoveContainer" containerID="d0ac0e16d4b8d3d3d2549819a456510f82fe5c4e2386c11d93e9431c732b82d7" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.590841 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:01:27 crc kubenswrapper[4559]: E1123 07:01:27.591287 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b00137d1-5df1-4e95-b066-827c4d2d53f8" containerName="extract-utilities" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.591305 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="b00137d1-5df1-4e95-b066-827c4d2d53f8" containerName="extract-utilities" Nov 23 07:01:27 crc kubenswrapper[4559]: E1123 07:01:27.591322 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b00137d1-5df1-4e95-b066-827c4d2d53f8" containerName="registry-server" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.591328 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="b00137d1-5df1-4e95-b066-827c4d2d53f8" containerName="registry-server" Nov 23 07:01:27 crc kubenswrapper[4559]: E1123 07:01:27.591359 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b00137d1-5df1-4e95-b066-827c4d2d53f8" containerName="extract-content" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.591365 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="b00137d1-5df1-4e95-b066-827c4d2d53f8" containerName="extract-content" Nov 23 07:01:27 crc kubenswrapper[4559]: E1123 07:01:27.591374 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2522249b-6744-4b19-9ffd-d5dbad83a1bd" containerName="nova-scheduler-scheduler" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.591382 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="2522249b-6744-4b19-9ffd-d5dbad83a1bd" containerName="nova-scheduler-scheduler" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.591622 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="2522249b-6744-4b19-9ffd-d5dbad83a1bd" containerName="nova-scheduler-scheduler" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.591697 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="b00137d1-5df1-4e95-b066-827c4d2d53f8" containerName="registry-server" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.592296 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.595320 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.597251 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.610706 4559 scope.go:117] "RemoveContainer" containerID="56272bdfc95aa42872687268934b43ac5627b3a728fa2921f39a8e17eefc381f" Nov 23 07:01:27 crc kubenswrapper[4559]: E1123 07:01:27.610995 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56272bdfc95aa42872687268934b43ac5627b3a728fa2921f39a8e17eefc381f\": container with ID starting with 56272bdfc95aa42872687268934b43ac5627b3a728fa2921f39a8e17eefc381f not found: ID does not exist" containerID="56272bdfc95aa42872687268934b43ac5627b3a728fa2921f39a8e17eefc381f" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.611024 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56272bdfc95aa42872687268934b43ac5627b3a728fa2921f39a8e17eefc381f"} err="failed to get container status \"56272bdfc95aa42872687268934b43ac5627b3a728fa2921f39a8e17eefc381f\": rpc error: code = NotFound desc = could not find container \"56272bdfc95aa42872687268934b43ac5627b3a728fa2921f39a8e17eefc381f\": container with ID starting with 56272bdfc95aa42872687268934b43ac5627b3a728fa2921f39a8e17eefc381f not found: ID does not exist" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.611043 4559 scope.go:117] "RemoveContainer" containerID="b731d59a4e4db9698acd3df6e767a006b3057c8bc2bd86a16fa6cbaecdbc5797" Nov 23 07:01:27 crc kubenswrapper[4559]: E1123 07:01:27.611399 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b731d59a4e4db9698acd3df6e767a006b3057c8bc2bd86a16fa6cbaecdbc5797\": container with ID starting with b731d59a4e4db9698acd3df6e767a006b3057c8bc2bd86a16fa6cbaecdbc5797 not found: ID does not exist" containerID="b731d59a4e4db9698acd3df6e767a006b3057c8bc2bd86a16fa6cbaecdbc5797" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.611418 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b731d59a4e4db9698acd3df6e767a006b3057c8bc2bd86a16fa6cbaecdbc5797"} err="failed to get container status \"b731d59a4e4db9698acd3df6e767a006b3057c8bc2bd86a16fa6cbaecdbc5797\": rpc error: code = NotFound desc = could not find container \"b731d59a4e4db9698acd3df6e767a006b3057c8bc2bd86a16fa6cbaecdbc5797\": container with ID starting with b731d59a4e4db9698acd3df6e767a006b3057c8bc2bd86a16fa6cbaecdbc5797 not found: ID does not exist" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.611430 4559 scope.go:117] "RemoveContainer" containerID="d0ac0e16d4b8d3d3d2549819a456510f82fe5c4e2386c11d93e9431c732b82d7" Nov 23 07:01:27 crc kubenswrapper[4559]: E1123 07:01:27.611681 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0ac0e16d4b8d3d3d2549819a456510f82fe5c4e2386c11d93e9431c732b82d7\": container with ID starting with d0ac0e16d4b8d3d3d2549819a456510f82fe5c4e2386c11d93e9431c732b82d7 not found: ID does not exist" containerID="d0ac0e16d4b8d3d3d2549819a456510f82fe5c4e2386c11d93e9431c732b82d7" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.611702 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0ac0e16d4b8d3d3d2549819a456510f82fe5c4e2386c11d93e9431c732b82d7"} err="failed to get container status \"d0ac0e16d4b8d3d3d2549819a456510f82fe5c4e2386c11d93e9431c732b82d7\": rpc error: code = NotFound desc = could not find container \"d0ac0e16d4b8d3d3d2549819a456510f82fe5c4e2386c11d93e9431c732b82d7\": container with ID starting with d0ac0e16d4b8d3d3d2549819a456510f82fe5c4e2386c11d93e9431c732b82d7 not found: ID does not exist" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.687024 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27bbb5a9-7e24-4529-a05e-1a0bdc5c7572-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"27bbb5a9-7e24-4529-a05e-1a0bdc5c7572\") " pod="openstack/nova-scheduler-0" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.687138 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27bbb5a9-7e24-4529-a05e-1a0bdc5c7572-config-data\") pod \"nova-scheduler-0\" (UID: \"27bbb5a9-7e24-4529-a05e-1a0bdc5c7572\") " pod="openstack/nova-scheduler-0" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.687175 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lswtn\" (UniqueName: \"kubernetes.io/projected/27bbb5a9-7e24-4529-a05e-1a0bdc5c7572-kube-api-access-lswtn\") pod \"nova-scheduler-0\" (UID: \"27bbb5a9-7e24-4529-a05e-1a0bdc5c7572\") " pod="openstack/nova-scheduler-0" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.789612 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27bbb5a9-7e24-4529-a05e-1a0bdc5c7572-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"27bbb5a9-7e24-4529-a05e-1a0bdc5c7572\") " pod="openstack/nova-scheduler-0" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.789836 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27bbb5a9-7e24-4529-a05e-1a0bdc5c7572-config-data\") pod \"nova-scheduler-0\" (UID: \"27bbb5a9-7e24-4529-a05e-1a0bdc5c7572\") " pod="openstack/nova-scheduler-0" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.789884 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lswtn\" (UniqueName: \"kubernetes.io/projected/27bbb5a9-7e24-4529-a05e-1a0bdc5c7572-kube-api-access-lswtn\") pod \"nova-scheduler-0\" (UID: \"27bbb5a9-7e24-4529-a05e-1a0bdc5c7572\") " pod="openstack/nova-scheduler-0" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.793090 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27bbb5a9-7e24-4529-a05e-1a0bdc5c7572-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"27bbb5a9-7e24-4529-a05e-1a0bdc5c7572\") " pod="openstack/nova-scheduler-0" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.793290 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27bbb5a9-7e24-4529-a05e-1a0bdc5c7572-config-data\") pod \"nova-scheduler-0\" (UID: \"27bbb5a9-7e24-4529-a05e-1a0bdc5c7572\") " pod="openstack/nova-scheduler-0" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.807368 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lswtn\" (UniqueName: \"kubernetes.io/projected/27bbb5a9-7e24-4529-a05e-1a0bdc5c7572-kube-api-access-lswtn\") pod \"nova-scheduler-0\" (UID: \"27bbb5a9-7e24-4529-a05e-1a0bdc5c7572\") " pod="openstack/nova-scheduler-0" Nov 23 07:01:27 crc kubenswrapper[4559]: I1123 07:01:27.905761 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 07:01:28 crc kubenswrapper[4559]: I1123 07:01:28.115386 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:01:28 crc kubenswrapper[4559]: W1123 07:01:28.119443 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27bbb5a9_7e24_4529_a05e_1a0bdc5c7572.slice/crio-9328cf35b5c593f87eebdf648275dde55e2b3b4cf8ba43cc8d1775bdd21c3f19 WatchSource:0}: Error finding container 9328cf35b5c593f87eebdf648275dde55e2b3b4cf8ba43cc8d1775bdd21c3f19: Status 404 returned error can't find the container with id 9328cf35b5c593f87eebdf648275dde55e2b3b4cf8ba43cc8d1775bdd21c3f19 Nov 23 07:01:28 crc kubenswrapper[4559]: I1123 07:01:28.287488 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2522249b-6744-4b19-9ffd-d5dbad83a1bd" path="/var/lib/kubelet/pods/2522249b-6744-4b19-9ffd-d5dbad83a1bd/volumes" Nov 23 07:01:28 crc kubenswrapper[4559]: I1123 07:01:28.288134 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b00137d1-5df1-4e95-b066-827c4d2d53f8" path="/var/lib/kubelet/pods/b00137d1-5df1-4e95-b066-827c4d2d53f8/volumes" Nov 23 07:01:28 crc kubenswrapper[4559]: I1123 07:01:28.530107 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"27bbb5a9-7e24-4529-a05e-1a0bdc5c7572","Type":"ContainerStarted","Data":"6eff3da2d979d6fd24ecae38d241fea26cc86e4a77ec0a6676a5d6e92fdc15a4"} Nov 23 07:01:28 crc kubenswrapper[4559]: I1123 07:01:28.530159 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"27bbb5a9-7e24-4529-a05e-1a0bdc5c7572","Type":"ContainerStarted","Data":"9328cf35b5c593f87eebdf648275dde55e2b3b4cf8ba43cc8d1775bdd21c3f19"} Nov 23 07:01:28 crc kubenswrapper[4559]: I1123 07:01:28.552281 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.552267823 podStartE2EDuration="1.552267823s" podCreationTimestamp="2025-11-23 07:01:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:01:28.547612879 +0000 UTC m=+990.569598494" watchObservedRunningTime="2025-11-23 07:01:28.552267823 +0000 UTC m=+990.574253437" Nov 23 07:01:28 crc kubenswrapper[4559]: I1123 07:01:28.933588 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 23 07:01:28 crc kubenswrapper[4559]: I1123 07:01:28.933671 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 23 07:01:29 crc kubenswrapper[4559]: I1123 07:01:29.765106 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 23 07:01:31 crc kubenswrapper[4559]: I1123 07:01:31.707487 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-99bwv" Nov 23 07:01:31 crc kubenswrapper[4559]: I1123 07:01:31.743026 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-99bwv" Nov 23 07:01:31 crc kubenswrapper[4559]: I1123 07:01:31.952357 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-99bwv"] Nov 23 07:01:32 crc kubenswrapper[4559]: I1123 07:01:32.906971 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 23 07:01:33 crc kubenswrapper[4559]: I1123 07:01:33.482567 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 23 07:01:33 crc kubenswrapper[4559]: I1123 07:01:33.578219 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-99bwv" podUID="63adda33-68e6-456d-beac-04903b136067" containerName="registry-server" containerID="cri-o://bb49ec902c3bb819a91dd78740a3390349874432ece305c9d9017e7dd67bcf65" gracePeriod=2 Nov 23 07:01:33 crc kubenswrapper[4559]: I1123 07:01:33.934218 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 23 07:01:33 crc kubenswrapper[4559]: I1123 07:01:33.934279 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 23 07:01:33 crc kubenswrapper[4559]: I1123 07:01:33.962927 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 23 07:01:33 crc kubenswrapper[4559]: I1123 07:01:33.963014 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 23 07:01:33 crc kubenswrapper[4559]: I1123 07:01:33.993473 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-99bwv" Nov 23 07:01:34 crc kubenswrapper[4559]: I1123 07:01:34.107464 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63adda33-68e6-456d-beac-04903b136067-utilities\") pod \"63adda33-68e6-456d-beac-04903b136067\" (UID: \"63adda33-68e6-456d-beac-04903b136067\") " Nov 23 07:01:34 crc kubenswrapper[4559]: I1123 07:01:34.107841 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgctz\" (UniqueName: \"kubernetes.io/projected/63adda33-68e6-456d-beac-04903b136067-kube-api-access-jgctz\") pod \"63adda33-68e6-456d-beac-04903b136067\" (UID: \"63adda33-68e6-456d-beac-04903b136067\") " Nov 23 07:01:34 crc kubenswrapper[4559]: I1123 07:01:34.108302 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63adda33-68e6-456d-beac-04903b136067-utilities" (OuterVolumeSpecName: "utilities") pod "63adda33-68e6-456d-beac-04903b136067" (UID: "63adda33-68e6-456d-beac-04903b136067"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:01:34 crc kubenswrapper[4559]: I1123 07:01:34.108921 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63adda33-68e6-456d-beac-04903b136067-catalog-content\") pod \"63adda33-68e6-456d-beac-04903b136067\" (UID: \"63adda33-68e6-456d-beac-04903b136067\") " Nov 23 07:01:34 crc kubenswrapper[4559]: I1123 07:01:34.109499 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63adda33-68e6-456d-beac-04903b136067-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:34 crc kubenswrapper[4559]: I1123 07:01:34.113430 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63adda33-68e6-456d-beac-04903b136067-kube-api-access-jgctz" (OuterVolumeSpecName: "kube-api-access-jgctz") pod "63adda33-68e6-456d-beac-04903b136067" (UID: "63adda33-68e6-456d-beac-04903b136067"). InnerVolumeSpecName "kube-api-access-jgctz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:01:34 crc kubenswrapper[4559]: I1123 07:01:34.190130 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63adda33-68e6-456d-beac-04903b136067-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "63adda33-68e6-456d-beac-04903b136067" (UID: "63adda33-68e6-456d-beac-04903b136067"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:01:34 crc kubenswrapper[4559]: I1123 07:01:34.212218 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgctz\" (UniqueName: \"kubernetes.io/projected/63adda33-68e6-456d-beac-04903b136067-kube-api-access-jgctz\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:34 crc kubenswrapper[4559]: I1123 07:01:34.212257 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63adda33-68e6-456d-beac-04903b136067-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:34 crc kubenswrapper[4559]: I1123 07:01:34.590855 4559 generic.go:334] "Generic (PLEG): container finished" podID="63adda33-68e6-456d-beac-04903b136067" containerID="bb49ec902c3bb819a91dd78740a3390349874432ece305c9d9017e7dd67bcf65" exitCode=0 Nov 23 07:01:34 crc kubenswrapper[4559]: I1123 07:01:34.590923 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-99bwv" event={"ID":"63adda33-68e6-456d-beac-04903b136067","Type":"ContainerDied","Data":"bb49ec902c3bb819a91dd78740a3390349874432ece305c9d9017e7dd67bcf65"} Nov 23 07:01:34 crc kubenswrapper[4559]: I1123 07:01:34.590962 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-99bwv" event={"ID":"63adda33-68e6-456d-beac-04903b136067","Type":"ContainerDied","Data":"1a024a75fb8fcdd87f2ddc5b536131955d9d4e2bc289ea795108fa56e6cba5b6"} Nov 23 07:01:34 crc kubenswrapper[4559]: I1123 07:01:34.590963 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-99bwv" Nov 23 07:01:34 crc kubenswrapper[4559]: I1123 07:01:34.590993 4559 scope.go:117] "RemoveContainer" containerID="bb49ec902c3bb819a91dd78740a3390349874432ece305c9d9017e7dd67bcf65" Nov 23 07:01:34 crc kubenswrapper[4559]: I1123 07:01:34.614860 4559 scope.go:117] "RemoveContainer" containerID="bb5a9af84f00a0714495fb3402d6db3ed49864b36c7ea8f94ebfa038cbfb4fd2" Nov 23 07:01:34 crc kubenswrapper[4559]: I1123 07:01:34.618421 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-99bwv"] Nov 23 07:01:34 crc kubenswrapper[4559]: I1123 07:01:34.627588 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-99bwv"] Nov 23 07:01:34 crc kubenswrapper[4559]: I1123 07:01:34.633906 4559 scope.go:117] "RemoveContainer" containerID="8d0c7c6c63cdc10376edbec504e38196ad23ef4d45e38ff36e5688eb8e55b3ef" Nov 23 07:01:34 crc kubenswrapper[4559]: I1123 07:01:34.665914 4559 scope.go:117] "RemoveContainer" containerID="bb49ec902c3bb819a91dd78740a3390349874432ece305c9d9017e7dd67bcf65" Nov 23 07:01:34 crc kubenswrapper[4559]: E1123 07:01:34.666469 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb49ec902c3bb819a91dd78740a3390349874432ece305c9d9017e7dd67bcf65\": container with ID starting with bb49ec902c3bb819a91dd78740a3390349874432ece305c9d9017e7dd67bcf65 not found: ID does not exist" containerID="bb49ec902c3bb819a91dd78740a3390349874432ece305c9d9017e7dd67bcf65" Nov 23 07:01:34 crc kubenswrapper[4559]: I1123 07:01:34.666546 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb49ec902c3bb819a91dd78740a3390349874432ece305c9d9017e7dd67bcf65"} err="failed to get container status \"bb49ec902c3bb819a91dd78740a3390349874432ece305c9d9017e7dd67bcf65\": rpc error: code = NotFound desc = could not find container \"bb49ec902c3bb819a91dd78740a3390349874432ece305c9d9017e7dd67bcf65\": container with ID starting with bb49ec902c3bb819a91dd78740a3390349874432ece305c9d9017e7dd67bcf65 not found: ID does not exist" Nov 23 07:01:34 crc kubenswrapper[4559]: I1123 07:01:34.666595 4559 scope.go:117] "RemoveContainer" containerID="bb5a9af84f00a0714495fb3402d6db3ed49864b36c7ea8f94ebfa038cbfb4fd2" Nov 23 07:01:34 crc kubenswrapper[4559]: E1123 07:01:34.667081 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb5a9af84f00a0714495fb3402d6db3ed49864b36c7ea8f94ebfa038cbfb4fd2\": container with ID starting with bb5a9af84f00a0714495fb3402d6db3ed49864b36c7ea8f94ebfa038cbfb4fd2 not found: ID does not exist" containerID="bb5a9af84f00a0714495fb3402d6db3ed49864b36c7ea8f94ebfa038cbfb4fd2" Nov 23 07:01:34 crc kubenswrapper[4559]: I1123 07:01:34.667122 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb5a9af84f00a0714495fb3402d6db3ed49864b36c7ea8f94ebfa038cbfb4fd2"} err="failed to get container status \"bb5a9af84f00a0714495fb3402d6db3ed49864b36c7ea8f94ebfa038cbfb4fd2\": rpc error: code = NotFound desc = could not find container \"bb5a9af84f00a0714495fb3402d6db3ed49864b36c7ea8f94ebfa038cbfb4fd2\": container with ID starting with bb5a9af84f00a0714495fb3402d6db3ed49864b36c7ea8f94ebfa038cbfb4fd2 not found: ID does not exist" Nov 23 07:01:34 crc kubenswrapper[4559]: I1123 07:01:34.667154 4559 scope.go:117] "RemoveContainer" containerID="8d0c7c6c63cdc10376edbec504e38196ad23ef4d45e38ff36e5688eb8e55b3ef" Nov 23 07:01:34 crc kubenswrapper[4559]: E1123 07:01:34.667532 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d0c7c6c63cdc10376edbec504e38196ad23ef4d45e38ff36e5688eb8e55b3ef\": container with ID starting with 8d0c7c6c63cdc10376edbec504e38196ad23ef4d45e38ff36e5688eb8e55b3ef not found: ID does not exist" containerID="8d0c7c6c63cdc10376edbec504e38196ad23ef4d45e38ff36e5688eb8e55b3ef" Nov 23 07:01:34 crc kubenswrapper[4559]: I1123 07:01:34.667557 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d0c7c6c63cdc10376edbec504e38196ad23ef4d45e38ff36e5688eb8e55b3ef"} err="failed to get container status \"8d0c7c6c63cdc10376edbec504e38196ad23ef4d45e38ff36e5688eb8e55b3ef\": rpc error: code = NotFound desc = could not find container \"8d0c7c6c63cdc10376edbec504e38196ad23ef4d45e38ff36e5688eb8e55b3ef\": container with ID starting with 8d0c7c6c63cdc10376edbec504e38196ad23ef4d45e38ff36e5688eb8e55b3ef not found: ID does not exist" Nov 23 07:01:34 crc kubenswrapper[4559]: I1123 07:01:34.952777 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8d0d4e86-3876-4666-b628-83cb6c08700b" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 23 07:01:34 crc kubenswrapper[4559]: I1123 07:01:34.952801 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8d0d4e86-3876-4666-b628-83cb6c08700b" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 23 07:01:35 crc kubenswrapper[4559]: I1123 07:01:35.045840 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7b0260a5-8273-4d39-a06f-b1d02ff9696e" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 23 07:01:35 crc kubenswrapper[4559]: I1123 07:01:35.045891 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7b0260a5-8273-4d39-a06f-b1d02ff9696e" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 23 07:01:36 crc kubenswrapper[4559]: I1123 07:01:36.286968 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63adda33-68e6-456d-beac-04903b136067" path="/var/lib/kubelet/pods/63adda33-68e6-456d-beac-04903b136067/volumes" Nov 23 07:01:36 crc kubenswrapper[4559]: I1123 07:01:36.883182 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 07:01:36 crc kubenswrapper[4559]: I1123 07:01:36.883723 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="f306bc71-8c65-4236-b439-e3af9c495a15" containerName="kube-state-metrics" containerID="cri-o://c75a09b9967ab3f8d926027629a6b224d1d1615027f5f57d8d16bb2c450be721" gracePeriod=30 Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.289218 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.482047 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nrt6\" (UniqueName: \"kubernetes.io/projected/f306bc71-8c65-4236-b439-e3af9c495a15-kube-api-access-4nrt6\") pod \"f306bc71-8c65-4236-b439-e3af9c495a15\" (UID: \"f306bc71-8c65-4236-b439-e3af9c495a15\") " Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.493870 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f306bc71-8c65-4236-b439-e3af9c495a15-kube-api-access-4nrt6" (OuterVolumeSpecName: "kube-api-access-4nrt6") pod "f306bc71-8c65-4236-b439-e3af9c495a15" (UID: "f306bc71-8c65-4236-b439-e3af9c495a15"). InnerVolumeSpecName "kube-api-access-4nrt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.584864 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nrt6\" (UniqueName: \"kubernetes.io/projected/f306bc71-8c65-4236-b439-e3af9c495a15-kube-api-access-4nrt6\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.619169 4559 generic.go:334] "Generic (PLEG): container finished" podID="f306bc71-8c65-4236-b439-e3af9c495a15" containerID="c75a09b9967ab3f8d926027629a6b224d1d1615027f5f57d8d16bb2c450be721" exitCode=2 Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.619275 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.619310 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f306bc71-8c65-4236-b439-e3af9c495a15","Type":"ContainerDied","Data":"c75a09b9967ab3f8d926027629a6b224d1d1615027f5f57d8d16bb2c450be721"} Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.619410 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f306bc71-8c65-4236-b439-e3af9c495a15","Type":"ContainerDied","Data":"a82e6e70536b9910866670d1d9955a3e45d2d9781f262ee685560c24e30f92ce"} Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.619441 4559 scope.go:117] "RemoveContainer" containerID="c75a09b9967ab3f8d926027629a6b224d1d1615027f5f57d8d16bb2c450be721" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.638247 4559 scope.go:117] "RemoveContainer" containerID="c75a09b9967ab3f8d926027629a6b224d1d1615027f5f57d8d16bb2c450be721" Nov 23 07:01:37 crc kubenswrapper[4559]: E1123 07:01:37.638569 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c75a09b9967ab3f8d926027629a6b224d1d1615027f5f57d8d16bb2c450be721\": container with ID starting with c75a09b9967ab3f8d926027629a6b224d1d1615027f5f57d8d16bb2c450be721 not found: ID does not exist" containerID="c75a09b9967ab3f8d926027629a6b224d1d1615027f5f57d8d16bb2c450be721" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.638603 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c75a09b9967ab3f8d926027629a6b224d1d1615027f5f57d8d16bb2c450be721"} err="failed to get container status \"c75a09b9967ab3f8d926027629a6b224d1d1615027f5f57d8d16bb2c450be721\": rpc error: code = NotFound desc = could not find container \"c75a09b9967ab3f8d926027629a6b224d1d1615027f5f57d8d16bb2c450be721\": container with ID starting with c75a09b9967ab3f8d926027629a6b224d1d1615027f5f57d8d16bb2c450be721 not found: ID does not exist" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.646830 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.652577 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.665049 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 07:01:37 crc kubenswrapper[4559]: E1123 07:01:37.665387 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63adda33-68e6-456d-beac-04903b136067" containerName="extract-utilities" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.665406 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="63adda33-68e6-456d-beac-04903b136067" containerName="extract-utilities" Nov 23 07:01:37 crc kubenswrapper[4559]: E1123 07:01:37.665414 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63adda33-68e6-456d-beac-04903b136067" containerName="registry-server" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.665420 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="63adda33-68e6-456d-beac-04903b136067" containerName="registry-server" Nov 23 07:01:37 crc kubenswrapper[4559]: E1123 07:01:37.665440 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f306bc71-8c65-4236-b439-e3af9c495a15" containerName="kube-state-metrics" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.665446 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="f306bc71-8c65-4236-b439-e3af9c495a15" containerName="kube-state-metrics" Nov 23 07:01:37 crc kubenswrapper[4559]: E1123 07:01:37.665460 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63adda33-68e6-456d-beac-04903b136067" containerName="extract-content" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.665465 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="63adda33-68e6-456d-beac-04903b136067" containerName="extract-content" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.665620 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="63adda33-68e6-456d-beac-04903b136067" containerName="registry-server" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.665636 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="f306bc71-8c65-4236-b439-e3af9c495a15" containerName="kube-state-metrics" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.666252 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.672085 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.672308 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.702444 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.794084 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbrnf\" (UniqueName: \"kubernetes.io/projected/916e4424-5f09-44c7-8b07-de2a4d84df18-kube-api-access-tbrnf\") pod \"kube-state-metrics-0\" (UID: \"916e4424-5f09-44c7-8b07-de2a4d84df18\") " pod="openstack/kube-state-metrics-0" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.794285 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/916e4424-5f09-44c7-8b07-de2a4d84df18-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"916e4424-5f09-44c7-8b07-de2a4d84df18\") " pod="openstack/kube-state-metrics-0" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.794321 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/916e4424-5f09-44c7-8b07-de2a4d84df18-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"916e4424-5f09-44c7-8b07-de2a4d84df18\") " pod="openstack/kube-state-metrics-0" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.794493 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/916e4424-5f09-44c7-8b07-de2a4d84df18-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"916e4424-5f09-44c7-8b07-de2a4d84df18\") " pod="openstack/kube-state-metrics-0" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.896422 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/916e4424-5f09-44c7-8b07-de2a4d84df18-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"916e4424-5f09-44c7-8b07-de2a4d84df18\") " pod="openstack/kube-state-metrics-0" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.896998 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbrnf\" (UniqueName: \"kubernetes.io/projected/916e4424-5f09-44c7-8b07-de2a4d84df18-kube-api-access-tbrnf\") pod \"kube-state-metrics-0\" (UID: \"916e4424-5f09-44c7-8b07-de2a4d84df18\") " pod="openstack/kube-state-metrics-0" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.897146 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/916e4424-5f09-44c7-8b07-de2a4d84df18-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"916e4424-5f09-44c7-8b07-de2a4d84df18\") " pod="openstack/kube-state-metrics-0" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.897192 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/916e4424-5f09-44c7-8b07-de2a4d84df18-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"916e4424-5f09-44c7-8b07-de2a4d84df18\") " pod="openstack/kube-state-metrics-0" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.901211 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/916e4424-5f09-44c7-8b07-de2a4d84df18-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"916e4424-5f09-44c7-8b07-de2a4d84df18\") " pod="openstack/kube-state-metrics-0" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.904139 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/916e4424-5f09-44c7-8b07-de2a4d84df18-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"916e4424-5f09-44c7-8b07-de2a4d84df18\") " pod="openstack/kube-state-metrics-0" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.905261 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/916e4424-5f09-44c7-8b07-de2a4d84df18-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"916e4424-5f09-44c7-8b07-de2a4d84df18\") " pod="openstack/kube-state-metrics-0" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.909283 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.915443 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbrnf\" (UniqueName: \"kubernetes.io/projected/916e4424-5f09-44c7-8b07-de2a4d84df18-kube-api-access-tbrnf\") pod \"kube-state-metrics-0\" (UID: \"916e4424-5f09-44c7-8b07-de2a4d84df18\") " pod="openstack/kube-state-metrics-0" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.932754 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 23 07:01:37 crc kubenswrapper[4559]: I1123 07:01:37.982763 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 23 07:01:38 crc kubenswrapper[4559]: I1123 07:01:38.283213 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f306bc71-8c65-4236-b439-e3af9c495a15" path="/var/lib/kubelet/pods/f306bc71-8c65-4236-b439-e3af9c495a15/volumes" Nov 23 07:01:38 crc kubenswrapper[4559]: I1123 07:01:38.414341 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 07:01:38 crc kubenswrapper[4559]: I1123 07:01:38.569803 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:01:38 crc kubenswrapper[4559]: I1123 07:01:38.570050 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e76b4aa1-64f3-480b-a134-5ac6f6c02f04" containerName="ceilometer-central-agent" containerID="cri-o://54e5ac38d914c27d936b52c687e86987b540816ad3e7956c3f4263888aac90b4" gracePeriod=30 Nov 23 07:01:38 crc kubenswrapper[4559]: I1123 07:01:38.570086 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e76b4aa1-64f3-480b-a134-5ac6f6c02f04" containerName="proxy-httpd" containerID="cri-o://f9044311486a20b6b19f917e0121c85bad3325fd50152023e39c848cc28da22d" gracePeriod=30 Nov 23 07:01:38 crc kubenswrapper[4559]: I1123 07:01:38.570145 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e76b4aa1-64f3-480b-a134-5ac6f6c02f04" containerName="sg-core" containerID="cri-o://4b08563b50b2eb7ce1a5a2e63a6d28fef78d8987d89051f15a197f44fe01d399" gracePeriod=30 Nov 23 07:01:38 crc kubenswrapper[4559]: I1123 07:01:38.570155 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e76b4aa1-64f3-480b-a134-5ac6f6c02f04" containerName="ceilometer-notification-agent" containerID="cri-o://85e3166a92990c3ebc8bb481e3392e6bcd5139c8f8518e7ca870262593a5b5ef" gracePeriod=30 Nov 23 07:01:38 crc kubenswrapper[4559]: I1123 07:01:38.627849 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"916e4424-5f09-44c7-8b07-de2a4d84df18","Type":"ContainerStarted","Data":"420afb8069b4dc337b63cebc23998ae9bae5ea8f5ea86bc7c495b5626fed57ac"} Nov 23 07:01:38 crc kubenswrapper[4559]: I1123 07:01:38.654005 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 23 07:01:39 crc kubenswrapper[4559]: I1123 07:01:39.639239 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"916e4424-5f09-44c7-8b07-de2a4d84df18","Type":"ContainerStarted","Data":"1cf4fd23b6937f58c58e3896d2f6d0bee760c131f95702a9c854b3c610e0cb9a"} Nov 23 07:01:39 crc kubenswrapper[4559]: I1123 07:01:39.639845 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 23 07:01:39 crc kubenswrapper[4559]: I1123 07:01:39.643993 4559 generic.go:334] "Generic (PLEG): container finished" podID="e76b4aa1-64f3-480b-a134-5ac6f6c02f04" containerID="f9044311486a20b6b19f917e0121c85bad3325fd50152023e39c848cc28da22d" exitCode=0 Nov 23 07:01:39 crc kubenswrapper[4559]: I1123 07:01:39.644018 4559 generic.go:334] "Generic (PLEG): container finished" podID="e76b4aa1-64f3-480b-a134-5ac6f6c02f04" containerID="4b08563b50b2eb7ce1a5a2e63a6d28fef78d8987d89051f15a197f44fe01d399" exitCode=2 Nov 23 07:01:39 crc kubenswrapper[4559]: I1123 07:01:39.644025 4559 generic.go:334] "Generic (PLEG): container finished" podID="e76b4aa1-64f3-480b-a134-5ac6f6c02f04" containerID="54e5ac38d914c27d936b52c687e86987b540816ad3e7956c3f4263888aac90b4" exitCode=0 Nov 23 07:01:39 crc kubenswrapper[4559]: I1123 07:01:39.644064 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e76b4aa1-64f3-480b-a134-5ac6f6c02f04","Type":"ContainerDied","Data":"f9044311486a20b6b19f917e0121c85bad3325fd50152023e39c848cc28da22d"} Nov 23 07:01:39 crc kubenswrapper[4559]: I1123 07:01:39.644092 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e76b4aa1-64f3-480b-a134-5ac6f6c02f04","Type":"ContainerDied","Data":"4b08563b50b2eb7ce1a5a2e63a6d28fef78d8987d89051f15a197f44fe01d399"} Nov 23 07:01:39 crc kubenswrapper[4559]: I1123 07:01:39.644103 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e76b4aa1-64f3-480b-a134-5ac6f6c02f04","Type":"ContainerDied","Data":"54e5ac38d914c27d936b52c687e86987b540816ad3e7956c3f4263888aac90b4"} Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.161609 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.183187 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.8881234449999997 podStartE2EDuration="4.183168654s" podCreationTimestamp="2025-11-23 07:01:37 +0000 UTC" firstStartedPulling="2025-11-23 07:01:38.42024467 +0000 UTC m=+1000.442230284" lastFinishedPulling="2025-11-23 07:01:38.715289889 +0000 UTC m=+1000.737275493" observedRunningTime="2025-11-23 07:01:39.657487706 +0000 UTC m=+1001.679473320" watchObservedRunningTime="2025-11-23 07:01:41.183168654 +0000 UTC m=+1003.205154268" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.260461 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-log-httpd\") pod \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\" (UID: \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\") " Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.260531 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-scripts\") pod \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\" (UID: \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\") " Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.260679 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-run-httpd\") pod \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\" (UID: \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\") " Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.260782 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glgpq\" (UniqueName: \"kubernetes.io/projected/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-kube-api-access-glgpq\") pod \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\" (UID: \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\") " Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.260816 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-combined-ca-bundle\") pod \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\" (UID: \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\") " Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.261225 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-config-data\") pod \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\" (UID: \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\") " Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.260913 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e76b4aa1-64f3-480b-a134-5ac6f6c02f04" (UID: "e76b4aa1-64f3-480b-a134-5ac6f6c02f04"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.261323 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-sg-core-conf-yaml\") pod \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\" (UID: \"e76b4aa1-64f3-480b-a134-5ac6f6c02f04\") " Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.261316 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e76b4aa1-64f3-480b-a134-5ac6f6c02f04" (UID: "e76b4aa1-64f3-480b-a134-5ac6f6c02f04"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.261799 4559 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.261811 4559 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.266379 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-kube-api-access-glgpq" (OuterVolumeSpecName: "kube-api-access-glgpq") pod "e76b4aa1-64f3-480b-a134-5ac6f6c02f04" (UID: "e76b4aa1-64f3-480b-a134-5ac6f6c02f04"). InnerVolumeSpecName "kube-api-access-glgpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.266546 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-scripts" (OuterVolumeSpecName: "scripts") pod "e76b4aa1-64f3-480b-a134-5ac6f6c02f04" (UID: "e76b4aa1-64f3-480b-a134-5ac6f6c02f04"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.286954 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e76b4aa1-64f3-480b-a134-5ac6f6c02f04" (UID: "e76b4aa1-64f3-480b-a134-5ac6f6c02f04"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.327686 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e76b4aa1-64f3-480b-a134-5ac6f6c02f04" (UID: "e76b4aa1-64f3-480b-a134-5ac6f6c02f04"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.340969 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-config-data" (OuterVolumeSpecName: "config-data") pod "e76b4aa1-64f3-480b-a134-5ac6f6c02f04" (UID: "e76b4aa1-64f3-480b-a134-5ac6f6c02f04"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.365289 4559 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.365323 4559 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.365336 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glgpq\" (UniqueName: \"kubernetes.io/projected/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-kube-api-access-glgpq\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.365352 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.365366 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e76b4aa1-64f3-480b-a134-5ac6f6c02f04-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.661097 4559 generic.go:334] "Generic (PLEG): container finished" podID="e76b4aa1-64f3-480b-a134-5ac6f6c02f04" containerID="85e3166a92990c3ebc8bb481e3392e6bcd5139c8f8518e7ca870262593a5b5ef" exitCode=0 Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.661139 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e76b4aa1-64f3-480b-a134-5ac6f6c02f04","Type":"ContainerDied","Data":"85e3166a92990c3ebc8bb481e3392e6bcd5139c8f8518e7ca870262593a5b5ef"} Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.661162 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.661182 4559 scope.go:117] "RemoveContainer" containerID="f9044311486a20b6b19f917e0121c85bad3325fd50152023e39c848cc28da22d" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.661171 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e76b4aa1-64f3-480b-a134-5ac6f6c02f04","Type":"ContainerDied","Data":"3d93e1f9bbb249f41a971a795b9e9ee7c19b6e1921e5be767ce24de7d4830925"} Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.677909 4559 scope.go:117] "RemoveContainer" containerID="4b08563b50b2eb7ce1a5a2e63a6d28fef78d8987d89051f15a197f44fe01d399" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.692136 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.695690 4559 scope.go:117] "RemoveContainer" containerID="85e3166a92990c3ebc8bb481e3392e6bcd5139c8f8518e7ca870262593a5b5ef" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.712730 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.720893 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:01:41 crc kubenswrapper[4559]: E1123 07:01:41.721377 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e76b4aa1-64f3-480b-a134-5ac6f6c02f04" containerName="ceilometer-notification-agent" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.721399 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="e76b4aa1-64f3-480b-a134-5ac6f6c02f04" containerName="ceilometer-notification-agent" Nov 23 07:01:41 crc kubenswrapper[4559]: E1123 07:01:41.721422 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e76b4aa1-64f3-480b-a134-5ac6f6c02f04" containerName="proxy-httpd" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.721429 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="e76b4aa1-64f3-480b-a134-5ac6f6c02f04" containerName="proxy-httpd" Nov 23 07:01:41 crc kubenswrapper[4559]: E1123 07:01:41.721451 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e76b4aa1-64f3-480b-a134-5ac6f6c02f04" containerName="sg-core" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.721843 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="e76b4aa1-64f3-480b-a134-5ac6f6c02f04" containerName="sg-core" Nov 23 07:01:41 crc kubenswrapper[4559]: E1123 07:01:41.722272 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e76b4aa1-64f3-480b-a134-5ac6f6c02f04" containerName="ceilometer-central-agent" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.722289 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="e76b4aa1-64f3-480b-a134-5ac6f6c02f04" containerName="ceilometer-central-agent" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.725739 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="e76b4aa1-64f3-480b-a134-5ac6f6c02f04" containerName="ceilometer-central-agent" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.725766 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="e76b4aa1-64f3-480b-a134-5ac6f6c02f04" containerName="proxy-httpd" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.725788 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="e76b4aa1-64f3-480b-a134-5ac6f6c02f04" containerName="ceilometer-notification-agent" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.725804 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="e76b4aa1-64f3-480b-a134-5ac6f6c02f04" containerName="sg-core" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.727886 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.732235 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.732441 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.733464 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.734152 4559 scope.go:117] "RemoveContainer" containerID="54e5ac38d914c27d936b52c687e86987b540816ad3e7956c3f4263888aac90b4" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.734692 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.797310 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c43876e-183f-49ad-9b4d-316f1d076bb2-log-httpd\") pod \"ceilometer-0\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " pod="openstack/ceilometer-0" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.797384 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " pod="openstack/ceilometer-0" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.797426 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-scripts\") pod \"ceilometer-0\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " pod="openstack/ceilometer-0" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.797449 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " pod="openstack/ceilometer-0" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.797555 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-config-data\") pod \"ceilometer-0\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " pod="openstack/ceilometer-0" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.797729 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwp2n\" (UniqueName: \"kubernetes.io/projected/8c43876e-183f-49ad-9b4d-316f1d076bb2-kube-api-access-mwp2n\") pod \"ceilometer-0\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " pod="openstack/ceilometer-0" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.802081 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c43876e-183f-49ad-9b4d-316f1d076bb2-run-httpd\") pod \"ceilometer-0\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " pod="openstack/ceilometer-0" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.802142 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " pod="openstack/ceilometer-0" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.802234 4559 scope.go:117] "RemoveContainer" containerID="f9044311486a20b6b19f917e0121c85bad3325fd50152023e39c848cc28da22d" Nov 23 07:01:41 crc kubenswrapper[4559]: E1123 07:01:41.818766 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9044311486a20b6b19f917e0121c85bad3325fd50152023e39c848cc28da22d\": container with ID starting with f9044311486a20b6b19f917e0121c85bad3325fd50152023e39c848cc28da22d not found: ID does not exist" containerID="f9044311486a20b6b19f917e0121c85bad3325fd50152023e39c848cc28da22d" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.818821 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9044311486a20b6b19f917e0121c85bad3325fd50152023e39c848cc28da22d"} err="failed to get container status \"f9044311486a20b6b19f917e0121c85bad3325fd50152023e39c848cc28da22d\": rpc error: code = NotFound desc = could not find container \"f9044311486a20b6b19f917e0121c85bad3325fd50152023e39c848cc28da22d\": container with ID starting with f9044311486a20b6b19f917e0121c85bad3325fd50152023e39c848cc28da22d not found: ID does not exist" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.818850 4559 scope.go:117] "RemoveContainer" containerID="4b08563b50b2eb7ce1a5a2e63a6d28fef78d8987d89051f15a197f44fe01d399" Nov 23 07:01:41 crc kubenswrapper[4559]: E1123 07:01:41.822730 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b08563b50b2eb7ce1a5a2e63a6d28fef78d8987d89051f15a197f44fe01d399\": container with ID starting with 4b08563b50b2eb7ce1a5a2e63a6d28fef78d8987d89051f15a197f44fe01d399 not found: ID does not exist" containerID="4b08563b50b2eb7ce1a5a2e63a6d28fef78d8987d89051f15a197f44fe01d399" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.822756 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b08563b50b2eb7ce1a5a2e63a6d28fef78d8987d89051f15a197f44fe01d399"} err="failed to get container status \"4b08563b50b2eb7ce1a5a2e63a6d28fef78d8987d89051f15a197f44fe01d399\": rpc error: code = NotFound desc = could not find container \"4b08563b50b2eb7ce1a5a2e63a6d28fef78d8987d89051f15a197f44fe01d399\": container with ID starting with 4b08563b50b2eb7ce1a5a2e63a6d28fef78d8987d89051f15a197f44fe01d399 not found: ID does not exist" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.822777 4559 scope.go:117] "RemoveContainer" containerID="85e3166a92990c3ebc8bb481e3392e6bcd5139c8f8518e7ca870262593a5b5ef" Nov 23 07:01:41 crc kubenswrapper[4559]: E1123 07:01:41.826716 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85e3166a92990c3ebc8bb481e3392e6bcd5139c8f8518e7ca870262593a5b5ef\": container with ID starting with 85e3166a92990c3ebc8bb481e3392e6bcd5139c8f8518e7ca870262593a5b5ef not found: ID does not exist" containerID="85e3166a92990c3ebc8bb481e3392e6bcd5139c8f8518e7ca870262593a5b5ef" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.826750 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85e3166a92990c3ebc8bb481e3392e6bcd5139c8f8518e7ca870262593a5b5ef"} err="failed to get container status \"85e3166a92990c3ebc8bb481e3392e6bcd5139c8f8518e7ca870262593a5b5ef\": rpc error: code = NotFound desc = could not find container \"85e3166a92990c3ebc8bb481e3392e6bcd5139c8f8518e7ca870262593a5b5ef\": container with ID starting with 85e3166a92990c3ebc8bb481e3392e6bcd5139c8f8518e7ca870262593a5b5ef not found: ID does not exist" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.826771 4559 scope.go:117] "RemoveContainer" containerID="54e5ac38d914c27d936b52c687e86987b540816ad3e7956c3f4263888aac90b4" Nov 23 07:01:41 crc kubenswrapper[4559]: E1123 07:01:41.830710 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54e5ac38d914c27d936b52c687e86987b540816ad3e7956c3f4263888aac90b4\": container with ID starting with 54e5ac38d914c27d936b52c687e86987b540816ad3e7956c3f4263888aac90b4 not found: ID does not exist" containerID="54e5ac38d914c27d936b52c687e86987b540816ad3e7956c3f4263888aac90b4" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.830734 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54e5ac38d914c27d936b52c687e86987b540816ad3e7956c3f4263888aac90b4"} err="failed to get container status \"54e5ac38d914c27d936b52c687e86987b540816ad3e7956c3f4263888aac90b4\": rpc error: code = NotFound desc = could not find container \"54e5ac38d914c27d936b52c687e86987b540816ad3e7956c3f4263888aac90b4\": container with ID starting with 54e5ac38d914c27d936b52c687e86987b540816ad3e7956c3f4263888aac90b4 not found: ID does not exist" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.904278 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-config-data\") pod \"ceilometer-0\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " pod="openstack/ceilometer-0" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.904399 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwp2n\" (UniqueName: \"kubernetes.io/projected/8c43876e-183f-49ad-9b4d-316f1d076bb2-kube-api-access-mwp2n\") pod \"ceilometer-0\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " pod="openstack/ceilometer-0" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.904437 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " pod="openstack/ceilometer-0" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.904451 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c43876e-183f-49ad-9b4d-316f1d076bb2-run-httpd\") pod \"ceilometer-0\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " pod="openstack/ceilometer-0" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.904617 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c43876e-183f-49ad-9b4d-316f1d076bb2-log-httpd\") pod \"ceilometer-0\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " pod="openstack/ceilometer-0" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.904691 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " pod="openstack/ceilometer-0" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.904739 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-scripts\") pod \"ceilometer-0\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " pod="openstack/ceilometer-0" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.904767 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " pod="openstack/ceilometer-0" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.905324 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c43876e-183f-49ad-9b4d-316f1d076bb2-log-httpd\") pod \"ceilometer-0\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " pod="openstack/ceilometer-0" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.905818 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c43876e-183f-49ad-9b4d-316f1d076bb2-run-httpd\") pod \"ceilometer-0\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " pod="openstack/ceilometer-0" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.909276 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-config-data\") pod \"ceilometer-0\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " pod="openstack/ceilometer-0" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.909654 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " pod="openstack/ceilometer-0" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.911680 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-scripts\") pod \"ceilometer-0\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " pod="openstack/ceilometer-0" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.911866 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " pod="openstack/ceilometer-0" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.913262 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " pod="openstack/ceilometer-0" Nov 23 07:01:41 crc kubenswrapper[4559]: I1123 07:01:41.919896 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwp2n\" (UniqueName: \"kubernetes.io/projected/8c43876e-183f-49ad-9b4d-316f1d076bb2-kube-api-access-mwp2n\") pod \"ceilometer-0\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " pod="openstack/ceilometer-0" Nov 23 07:01:42 crc kubenswrapper[4559]: I1123 07:01:42.059528 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:01:42 crc kubenswrapper[4559]: I1123 07:01:42.282246 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e76b4aa1-64f3-480b-a134-5ac6f6c02f04" path="/var/lib/kubelet/pods/e76b4aa1-64f3-480b-a134-5ac6f6c02f04/volumes" Nov 23 07:01:42 crc kubenswrapper[4559]: I1123 07:01:42.468964 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:01:42 crc kubenswrapper[4559]: W1123 07:01:42.472120 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c43876e_183f_49ad_9b4d_316f1d076bb2.slice/crio-c77bd2cd04fe1c20ccb7498adbb3151388f5a11c5ff8388f0397df2229328862 WatchSource:0}: Error finding container c77bd2cd04fe1c20ccb7498adbb3151388f5a11c5ff8388f0397df2229328862: Status 404 returned error can't find the container with id c77bd2cd04fe1c20ccb7498adbb3151388f5a11c5ff8388f0397df2229328862 Nov 23 07:01:42 crc kubenswrapper[4559]: I1123 07:01:42.669932 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c43876e-183f-49ad-9b4d-316f1d076bb2","Type":"ContainerStarted","Data":"c77bd2cd04fe1c20ccb7498adbb3151388f5a11c5ff8388f0397df2229328862"} Nov 23 07:01:43 crc kubenswrapper[4559]: I1123 07:01:43.681260 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c43876e-183f-49ad-9b4d-316f1d076bb2","Type":"ContainerStarted","Data":"c66de58ae6923b9aee034fc32e7b6b9a90dc9e25570679035586a68f97d2c804"} Nov 23 07:01:43 crc kubenswrapper[4559]: I1123 07:01:43.939564 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 23 07:01:43 crc kubenswrapper[4559]: I1123 07:01:43.940001 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 23 07:01:43 crc kubenswrapper[4559]: I1123 07:01:43.947526 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 23 07:01:43 crc kubenswrapper[4559]: I1123 07:01:43.970709 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 23 07:01:43 crc kubenswrapper[4559]: I1123 07:01:43.971115 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 23 07:01:43 crc kubenswrapper[4559]: I1123 07:01:43.974765 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 23 07:01:43 crc kubenswrapper[4559]: I1123 07:01:43.981737 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 23 07:01:44 crc kubenswrapper[4559]: I1123 07:01:44.691357 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c43876e-183f-49ad-9b4d-316f1d076bb2","Type":"ContainerStarted","Data":"a7392d9b52b44c2422c67927827e2f5d5cc98c6e01f107095bd6923a68020b98"} Nov 23 07:01:44 crc kubenswrapper[4559]: I1123 07:01:44.691887 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 23 07:01:44 crc kubenswrapper[4559]: I1123 07:01:44.695598 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 23 07:01:44 crc kubenswrapper[4559]: I1123 07:01:44.696699 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 23 07:01:44 crc kubenswrapper[4559]: I1123 07:01:44.868469 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-858b54d4df-vf9p9"] Nov 23 07:01:44 crc kubenswrapper[4559]: I1123 07:01:44.870546 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" Nov 23 07:01:44 crc kubenswrapper[4559]: I1123 07:01:44.905797 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-858b54d4df-vf9p9"] Nov 23 07:01:44 crc kubenswrapper[4559]: I1123 07:01:44.965194 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-dns-svc\") pod \"dnsmasq-dns-858b54d4df-vf9p9\" (UID: \"92077d58-8565-447b-839d-f830613c4f99\") " pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" Nov 23 07:01:44 crc kubenswrapper[4559]: I1123 07:01:44.965309 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-dns-swift-storage-0\") pod \"dnsmasq-dns-858b54d4df-vf9p9\" (UID: \"92077d58-8565-447b-839d-f830613c4f99\") " pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" Nov 23 07:01:44 crc kubenswrapper[4559]: I1123 07:01:44.965345 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgt9d\" (UniqueName: \"kubernetes.io/projected/92077d58-8565-447b-839d-f830613c4f99-kube-api-access-jgt9d\") pod \"dnsmasq-dns-858b54d4df-vf9p9\" (UID: \"92077d58-8565-447b-839d-f830613c4f99\") " pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" Nov 23 07:01:44 crc kubenswrapper[4559]: I1123 07:01:44.965391 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-ovsdbserver-sb\") pod \"dnsmasq-dns-858b54d4df-vf9p9\" (UID: \"92077d58-8565-447b-839d-f830613c4f99\") " pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" Nov 23 07:01:44 crc kubenswrapper[4559]: I1123 07:01:44.965439 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-config\") pod \"dnsmasq-dns-858b54d4df-vf9p9\" (UID: \"92077d58-8565-447b-839d-f830613c4f99\") " pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" Nov 23 07:01:44 crc kubenswrapper[4559]: I1123 07:01:44.965554 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-ovsdbserver-nb\") pod \"dnsmasq-dns-858b54d4df-vf9p9\" (UID: \"92077d58-8565-447b-839d-f830613c4f99\") " pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" Nov 23 07:01:45 crc kubenswrapper[4559]: I1123 07:01:45.067037 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-config\") pod \"dnsmasq-dns-858b54d4df-vf9p9\" (UID: \"92077d58-8565-447b-839d-f830613c4f99\") " pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" Nov 23 07:01:45 crc kubenswrapper[4559]: I1123 07:01:45.067240 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-ovsdbserver-nb\") pod \"dnsmasq-dns-858b54d4df-vf9p9\" (UID: \"92077d58-8565-447b-839d-f830613c4f99\") " pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" Nov 23 07:01:45 crc kubenswrapper[4559]: I1123 07:01:45.067381 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-dns-svc\") pod \"dnsmasq-dns-858b54d4df-vf9p9\" (UID: \"92077d58-8565-447b-839d-f830613c4f99\") " pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" Nov 23 07:01:45 crc kubenswrapper[4559]: I1123 07:01:45.067497 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-dns-swift-storage-0\") pod \"dnsmasq-dns-858b54d4df-vf9p9\" (UID: \"92077d58-8565-447b-839d-f830613c4f99\") " pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" Nov 23 07:01:45 crc kubenswrapper[4559]: I1123 07:01:45.067540 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgt9d\" (UniqueName: \"kubernetes.io/projected/92077d58-8565-447b-839d-f830613c4f99-kube-api-access-jgt9d\") pod \"dnsmasq-dns-858b54d4df-vf9p9\" (UID: \"92077d58-8565-447b-839d-f830613c4f99\") " pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" Nov 23 07:01:45 crc kubenswrapper[4559]: I1123 07:01:45.067601 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-ovsdbserver-sb\") pod \"dnsmasq-dns-858b54d4df-vf9p9\" (UID: \"92077d58-8565-447b-839d-f830613c4f99\") " pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" Nov 23 07:01:45 crc kubenswrapper[4559]: I1123 07:01:45.068158 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-config\") pod \"dnsmasq-dns-858b54d4df-vf9p9\" (UID: \"92077d58-8565-447b-839d-f830613c4f99\") " pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" Nov 23 07:01:45 crc kubenswrapper[4559]: I1123 07:01:45.068180 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-ovsdbserver-nb\") pod \"dnsmasq-dns-858b54d4df-vf9p9\" (UID: \"92077d58-8565-447b-839d-f830613c4f99\") " pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" Nov 23 07:01:45 crc kubenswrapper[4559]: I1123 07:01:45.068772 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-dns-swift-storage-0\") pod \"dnsmasq-dns-858b54d4df-vf9p9\" (UID: \"92077d58-8565-447b-839d-f830613c4f99\") " pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" Nov 23 07:01:45 crc kubenswrapper[4559]: I1123 07:01:45.069007 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-dns-svc\") pod \"dnsmasq-dns-858b54d4df-vf9p9\" (UID: \"92077d58-8565-447b-839d-f830613c4f99\") " pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" Nov 23 07:01:45 crc kubenswrapper[4559]: I1123 07:01:45.069606 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-ovsdbserver-sb\") pod \"dnsmasq-dns-858b54d4df-vf9p9\" (UID: \"92077d58-8565-447b-839d-f830613c4f99\") " pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" Nov 23 07:01:45 crc kubenswrapper[4559]: I1123 07:01:45.086942 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgt9d\" (UniqueName: \"kubernetes.io/projected/92077d58-8565-447b-839d-f830613c4f99-kube-api-access-jgt9d\") pod \"dnsmasq-dns-858b54d4df-vf9p9\" (UID: \"92077d58-8565-447b-839d-f830613c4f99\") " pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" Nov 23 07:01:45 crc kubenswrapper[4559]: I1123 07:01:45.187901 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" Nov 23 07:01:45 crc kubenswrapper[4559]: I1123 07:01:45.689058 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-858b54d4df-vf9p9"] Nov 23 07:01:45 crc kubenswrapper[4559]: I1123 07:01:45.702299 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" event={"ID":"92077d58-8565-447b-839d-f830613c4f99","Type":"ContainerStarted","Data":"b3d91a0cc7169fd6a3a85da207e9da9fc523ff99c70c2f1f3909d1cf27293ae8"} Nov 23 07:01:45 crc kubenswrapper[4559]: I1123 07:01:45.705100 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c43876e-183f-49ad-9b4d-316f1d076bb2","Type":"ContainerStarted","Data":"6797922434b536426e5ec7f1b846712bddf13cb8f90832b7e9fcdf30de645cea"} Nov 23 07:01:46 crc kubenswrapper[4559]: I1123 07:01:46.713263 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:46 crc kubenswrapper[4559]: I1123 07:01:46.716026 4559 generic.go:334] "Generic (PLEG): container finished" podID="bb7a9144-0988-44cc-9596-0cdf01b5abb3" containerID="37aff7304a6fa691f7155c46e4294d8257e6eaa652671e24f44cb8c31b40f0fe" exitCode=137 Nov 23 07:01:46 crc kubenswrapper[4559]: I1123 07:01:46.716090 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:46 crc kubenswrapper[4559]: I1123 07:01:46.716110 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bb7a9144-0988-44cc-9596-0cdf01b5abb3","Type":"ContainerDied","Data":"37aff7304a6fa691f7155c46e4294d8257e6eaa652671e24f44cb8c31b40f0fe"} Nov 23 07:01:46 crc kubenswrapper[4559]: I1123 07:01:46.716141 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bb7a9144-0988-44cc-9596-0cdf01b5abb3","Type":"ContainerDied","Data":"e2ac0b4dc977aef988123030922249a9b4cbd6101644ad736ac41e2a268cb2ed"} Nov 23 07:01:46 crc kubenswrapper[4559]: I1123 07:01:46.716161 4559 scope.go:117] "RemoveContainer" containerID="37aff7304a6fa691f7155c46e4294d8257e6eaa652671e24f44cb8c31b40f0fe" Nov 23 07:01:46 crc kubenswrapper[4559]: I1123 07:01:46.718379 4559 generic.go:334] "Generic (PLEG): container finished" podID="92077d58-8565-447b-839d-f830613c4f99" containerID="917a01baee638d05e9c0544a5b67da3dd33d8892fbc2c266bffb1f7352ffff34" exitCode=0 Nov 23 07:01:46 crc kubenswrapper[4559]: I1123 07:01:46.718489 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" event={"ID":"92077d58-8565-447b-839d-f830613c4f99","Type":"ContainerDied","Data":"917a01baee638d05e9c0544a5b67da3dd33d8892fbc2c266bffb1f7352ffff34"} Nov 23 07:01:46 crc kubenswrapper[4559]: I1123 07:01:46.721763 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c43876e-183f-49ad-9b4d-316f1d076bb2","Type":"ContainerStarted","Data":"52784fa03ce3137011aaef3af96c67e0238d5f167260baafb042a2c088f5aa38"} Nov 23 07:01:46 crc kubenswrapper[4559]: I1123 07:01:46.723150 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 23 07:01:46 crc kubenswrapper[4559]: I1123 07:01:46.736743 4559 scope.go:117] "RemoveContainer" containerID="37aff7304a6fa691f7155c46e4294d8257e6eaa652671e24f44cb8c31b40f0fe" Nov 23 07:01:46 crc kubenswrapper[4559]: E1123 07:01:46.738164 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37aff7304a6fa691f7155c46e4294d8257e6eaa652671e24f44cb8c31b40f0fe\": container with ID starting with 37aff7304a6fa691f7155c46e4294d8257e6eaa652671e24f44cb8c31b40f0fe not found: ID does not exist" containerID="37aff7304a6fa691f7155c46e4294d8257e6eaa652671e24f44cb8c31b40f0fe" Nov 23 07:01:46 crc kubenswrapper[4559]: I1123 07:01:46.738195 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37aff7304a6fa691f7155c46e4294d8257e6eaa652671e24f44cb8c31b40f0fe"} err="failed to get container status \"37aff7304a6fa691f7155c46e4294d8257e6eaa652671e24f44cb8c31b40f0fe\": rpc error: code = NotFound desc = could not find container \"37aff7304a6fa691f7155c46e4294d8257e6eaa652671e24f44cb8c31b40f0fe\": container with ID starting with 37aff7304a6fa691f7155c46e4294d8257e6eaa652671e24f44cb8c31b40f0fe not found: ID does not exist" Nov 23 07:01:46 crc kubenswrapper[4559]: I1123 07:01:46.791239 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.175217284 podStartE2EDuration="5.79122275s" podCreationTimestamp="2025-11-23 07:01:41 +0000 UTC" firstStartedPulling="2025-11-23 07:01:42.47548437 +0000 UTC m=+1004.497469984" lastFinishedPulling="2025-11-23 07:01:46.091489836 +0000 UTC m=+1008.113475450" observedRunningTime="2025-11-23 07:01:46.786485089 +0000 UTC m=+1008.808470703" watchObservedRunningTime="2025-11-23 07:01:46.79122275 +0000 UTC m=+1008.813208364" Nov 23 07:01:46 crc kubenswrapper[4559]: I1123 07:01:46.814076 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7a9144-0988-44cc-9596-0cdf01b5abb3-combined-ca-bundle\") pod \"bb7a9144-0988-44cc-9596-0cdf01b5abb3\" (UID: \"bb7a9144-0988-44cc-9596-0cdf01b5abb3\") " Nov 23 07:01:46 crc kubenswrapper[4559]: I1123 07:01:46.814142 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wlcgm\" (UniqueName: \"kubernetes.io/projected/bb7a9144-0988-44cc-9596-0cdf01b5abb3-kube-api-access-wlcgm\") pod \"bb7a9144-0988-44cc-9596-0cdf01b5abb3\" (UID: \"bb7a9144-0988-44cc-9596-0cdf01b5abb3\") " Nov 23 07:01:46 crc kubenswrapper[4559]: I1123 07:01:46.814287 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb7a9144-0988-44cc-9596-0cdf01b5abb3-config-data\") pod \"bb7a9144-0988-44cc-9596-0cdf01b5abb3\" (UID: \"bb7a9144-0988-44cc-9596-0cdf01b5abb3\") " Nov 23 07:01:46 crc kubenswrapper[4559]: I1123 07:01:46.832507 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb7a9144-0988-44cc-9596-0cdf01b5abb3-kube-api-access-wlcgm" (OuterVolumeSpecName: "kube-api-access-wlcgm") pod "bb7a9144-0988-44cc-9596-0cdf01b5abb3" (UID: "bb7a9144-0988-44cc-9596-0cdf01b5abb3"). InnerVolumeSpecName "kube-api-access-wlcgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:01:46 crc kubenswrapper[4559]: I1123 07:01:46.835422 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:01:46 crc kubenswrapper[4559]: I1123 07:01:46.841291 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb7a9144-0988-44cc-9596-0cdf01b5abb3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb7a9144-0988-44cc-9596-0cdf01b5abb3" (UID: "bb7a9144-0988-44cc-9596-0cdf01b5abb3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:46 crc kubenswrapper[4559]: I1123 07:01:46.844636 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb7a9144-0988-44cc-9596-0cdf01b5abb3-config-data" (OuterVolumeSpecName: "config-data") pod "bb7a9144-0988-44cc-9596-0cdf01b5abb3" (UID: "bb7a9144-0988-44cc-9596-0cdf01b5abb3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:46 crc kubenswrapper[4559]: I1123 07:01:46.916687 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7a9144-0988-44cc-9596-0cdf01b5abb3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:46 crc kubenswrapper[4559]: I1123 07:01:46.916713 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wlcgm\" (UniqueName: \"kubernetes.io/projected/bb7a9144-0988-44cc-9596-0cdf01b5abb3-kube-api-access-wlcgm\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:46 crc kubenswrapper[4559]: I1123 07:01:46.916743 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb7a9144-0988-44cc-9596-0cdf01b5abb3-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.051429 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.058762 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.065839 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.080477 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 07:01:47 crc kubenswrapper[4559]: E1123 07:01:47.080916 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb7a9144-0988-44cc-9596-0cdf01b5abb3" containerName="nova-cell1-novncproxy-novncproxy" Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.080936 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb7a9144-0988-44cc-9596-0cdf01b5abb3" containerName="nova-cell1-novncproxy-novncproxy" Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.081149 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb7a9144-0988-44cc-9596-0cdf01b5abb3" containerName="nova-cell1-novncproxy-novncproxy" Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.081797 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.084522 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.089135 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.089359 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.106775 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.121339 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.121444 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.121478 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cldr\" (UniqueName: \"kubernetes.io/projected/75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b-kube-api-access-5cldr\") pod \"nova-cell1-novncproxy-0\" (UID: \"75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.121561 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.121743 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.224353 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.224465 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.224490 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cldr\" (UniqueName: \"kubernetes.io/projected/75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b-kube-api-access-5cldr\") pod \"nova-cell1-novncproxy-0\" (UID: \"75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.224569 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.224708 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.231125 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.234522 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.234699 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.235298 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.240412 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cldr\" (UniqueName: \"kubernetes.io/projected/75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b-kube-api-access-5cldr\") pod \"nova-cell1-novncproxy-0\" (UID: \"75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.405200 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.742630 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7b0260a5-8273-4d39-a06f-b1d02ff9696e" containerName="nova-api-log" containerID="cri-o://75e3395816a677c9b178c87af51011cedca02a3b7d307f0a6cb1e864de5b6886" gracePeriod=30 Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.742874 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7b0260a5-8273-4d39-a06f-b1d02ff9696e" containerName="nova-api-api" containerID="cri-o://2a5677e783c8afdf414c7343add7bd352ebead848a4aff47acdacb3119347ebe" gracePeriod=30 Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.743229 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" event={"ID":"92077d58-8565-447b-839d-f830613c4f99","Type":"ContainerStarted","Data":"26cfd11e9113e63055219afc73e75d789ebbf82df81a54d2363d4b223a062e83"} Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.743275 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.779803 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" podStartSLOduration=3.779777035 podStartE2EDuration="3.779777035s" podCreationTimestamp="2025-11-23 07:01:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:01:47.762704789 +0000 UTC m=+1009.784690403" watchObservedRunningTime="2025-11-23 07:01:47.779777035 +0000 UTC m=+1009.801762649" Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.825695 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 07:01:47 crc kubenswrapper[4559]: W1123 07:01:47.830168 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75fcd1aa_df9b_4cc4_91aa_dde8e919ff9b.slice/crio-459f34409d5d7efb414551de94779a8df229f5fa6fd73a70ea1b1aa7ae9920ab WatchSource:0}: Error finding container 459f34409d5d7efb414551de94779a8df229f5fa6fd73a70ea1b1aa7ae9920ab: Status 404 returned error can't find the container with id 459f34409d5d7efb414551de94779a8df229f5fa6fd73a70ea1b1aa7ae9920ab Nov 23 07:01:47 crc kubenswrapper[4559]: I1123 07:01:47.992001 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 23 07:01:48 crc kubenswrapper[4559]: I1123 07:01:48.287002 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb7a9144-0988-44cc-9596-0cdf01b5abb3" path="/var/lib/kubelet/pods/bb7a9144-0988-44cc-9596-0cdf01b5abb3/volumes" Nov 23 07:01:48 crc kubenswrapper[4559]: I1123 07:01:48.752525 4559 generic.go:334] "Generic (PLEG): container finished" podID="7b0260a5-8273-4d39-a06f-b1d02ff9696e" containerID="75e3395816a677c9b178c87af51011cedca02a3b7d307f0a6cb1e864de5b6886" exitCode=143 Nov 23 07:01:48 crc kubenswrapper[4559]: I1123 07:01:48.752588 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7b0260a5-8273-4d39-a06f-b1d02ff9696e","Type":"ContainerDied","Data":"75e3395816a677c9b178c87af51011cedca02a3b7d307f0a6cb1e864de5b6886"} Nov 23 07:01:48 crc kubenswrapper[4559]: I1123 07:01:48.757953 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b","Type":"ContainerStarted","Data":"3e6c3733dfbeb83f175e3af2c5a0bd3ec24ac26e822d99021e91fc8fd191b64e"} Nov 23 07:01:48 crc kubenswrapper[4559]: I1123 07:01:48.757998 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b","Type":"ContainerStarted","Data":"459f34409d5d7efb414551de94779a8df229f5fa6fd73a70ea1b1aa7ae9920ab"} Nov 23 07:01:48 crc kubenswrapper[4559]: I1123 07:01:48.758220 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8c43876e-183f-49ad-9b4d-316f1d076bb2" containerName="ceilometer-central-agent" containerID="cri-o://c66de58ae6923b9aee034fc32e7b6b9a90dc9e25570679035586a68f97d2c804" gracePeriod=30 Nov 23 07:01:48 crc kubenswrapper[4559]: I1123 07:01:48.759174 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8c43876e-183f-49ad-9b4d-316f1d076bb2" containerName="proxy-httpd" containerID="cri-o://52784fa03ce3137011aaef3af96c67e0238d5f167260baafb042a2c088f5aa38" gracePeriod=30 Nov 23 07:01:48 crc kubenswrapper[4559]: I1123 07:01:48.759321 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8c43876e-183f-49ad-9b4d-316f1d076bb2" containerName="sg-core" containerID="cri-o://6797922434b536426e5ec7f1b846712bddf13cb8f90832b7e9fcdf30de645cea" gracePeriod=30 Nov 23 07:01:48 crc kubenswrapper[4559]: I1123 07:01:48.759396 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8c43876e-183f-49ad-9b4d-316f1d076bb2" containerName="ceilometer-notification-agent" containerID="cri-o://a7392d9b52b44c2422c67927827e2f5d5cc98c6e01f107095bd6923a68020b98" gracePeriod=30 Nov 23 07:01:48 crc kubenswrapper[4559]: I1123 07:01:48.771360 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=1.771342614 podStartE2EDuration="1.771342614s" podCreationTimestamp="2025-11-23 07:01:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:01:48.769578677 +0000 UTC m=+1010.791564290" watchObservedRunningTime="2025-11-23 07:01:48.771342614 +0000 UTC m=+1010.793328228" Nov 23 07:01:49 crc kubenswrapper[4559]: I1123 07:01:49.765850 4559 generic.go:334] "Generic (PLEG): container finished" podID="8c43876e-183f-49ad-9b4d-316f1d076bb2" containerID="52784fa03ce3137011aaef3af96c67e0238d5f167260baafb042a2c088f5aa38" exitCode=0 Nov 23 07:01:49 crc kubenswrapper[4559]: I1123 07:01:49.766138 4559 generic.go:334] "Generic (PLEG): container finished" podID="8c43876e-183f-49ad-9b4d-316f1d076bb2" containerID="6797922434b536426e5ec7f1b846712bddf13cb8f90832b7e9fcdf30de645cea" exitCode=2 Nov 23 07:01:49 crc kubenswrapper[4559]: I1123 07:01:49.766150 4559 generic.go:334] "Generic (PLEG): container finished" podID="8c43876e-183f-49ad-9b4d-316f1d076bb2" containerID="a7392d9b52b44c2422c67927827e2f5d5cc98c6e01f107095bd6923a68020b98" exitCode=0 Nov 23 07:01:49 crc kubenswrapper[4559]: I1123 07:01:49.766093 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c43876e-183f-49ad-9b4d-316f1d076bb2","Type":"ContainerDied","Data":"52784fa03ce3137011aaef3af96c67e0238d5f167260baafb042a2c088f5aa38"} Nov 23 07:01:49 crc kubenswrapper[4559]: I1123 07:01:49.766270 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c43876e-183f-49ad-9b4d-316f1d076bb2","Type":"ContainerDied","Data":"6797922434b536426e5ec7f1b846712bddf13cb8f90832b7e9fcdf30de645cea"} Nov 23 07:01:49 crc kubenswrapper[4559]: I1123 07:01:49.766307 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c43876e-183f-49ad-9b4d-316f1d076bb2","Type":"ContainerDied","Data":"a7392d9b52b44c2422c67927827e2f5d5cc98c6e01f107095bd6923a68020b98"} Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.089594 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.199882 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c43876e-183f-49ad-9b4d-316f1d076bb2-run-httpd\") pod \"8c43876e-183f-49ad-9b4d-316f1d076bb2\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.200246 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-config-data\") pod \"8c43876e-183f-49ad-9b4d-316f1d076bb2\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.200383 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-sg-core-conf-yaml\") pod \"8c43876e-183f-49ad-9b4d-316f1d076bb2\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.200496 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-scripts\") pod \"8c43876e-183f-49ad-9b4d-316f1d076bb2\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.200634 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c43876e-183f-49ad-9b4d-316f1d076bb2-log-httpd\") pod \"8c43876e-183f-49ad-9b4d-316f1d076bb2\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.200838 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwp2n\" (UniqueName: \"kubernetes.io/projected/8c43876e-183f-49ad-9b4d-316f1d076bb2-kube-api-access-mwp2n\") pod \"8c43876e-183f-49ad-9b4d-316f1d076bb2\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.201282 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-ceilometer-tls-certs\") pod \"8c43876e-183f-49ad-9b4d-316f1d076bb2\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.200419 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c43876e-183f-49ad-9b4d-316f1d076bb2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8c43876e-183f-49ad-9b4d-316f1d076bb2" (UID: "8c43876e-183f-49ad-9b4d-316f1d076bb2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.201388 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c43876e-183f-49ad-9b4d-316f1d076bb2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8c43876e-183f-49ad-9b4d-316f1d076bb2" (UID: "8c43876e-183f-49ad-9b4d-316f1d076bb2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.201952 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-combined-ca-bundle\") pod \"8c43876e-183f-49ad-9b4d-316f1d076bb2\" (UID: \"8c43876e-183f-49ad-9b4d-316f1d076bb2\") " Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.203061 4559 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c43876e-183f-49ad-9b4d-316f1d076bb2-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.203150 4559 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c43876e-183f-49ad-9b4d-316f1d076bb2-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.207154 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-scripts" (OuterVolumeSpecName: "scripts") pod "8c43876e-183f-49ad-9b4d-316f1d076bb2" (UID: "8c43876e-183f-49ad-9b4d-316f1d076bb2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.207433 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c43876e-183f-49ad-9b4d-316f1d076bb2-kube-api-access-mwp2n" (OuterVolumeSpecName: "kube-api-access-mwp2n") pod "8c43876e-183f-49ad-9b4d-316f1d076bb2" (UID: "8c43876e-183f-49ad-9b4d-316f1d076bb2"). InnerVolumeSpecName "kube-api-access-mwp2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.231613 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8c43876e-183f-49ad-9b4d-316f1d076bb2" (UID: "8c43876e-183f-49ad-9b4d-316f1d076bb2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.246524 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "8c43876e-183f-49ad-9b4d-316f1d076bb2" (UID: "8c43876e-183f-49ad-9b4d-316f1d076bb2"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.271626 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c43876e-183f-49ad-9b4d-316f1d076bb2" (UID: "8c43876e-183f-49ad-9b4d-316f1d076bb2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.286819 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-config-data" (OuterVolumeSpecName: "config-data") pod "8c43876e-183f-49ad-9b4d-316f1d076bb2" (UID: "8c43876e-183f-49ad-9b4d-316f1d076bb2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.305158 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.305226 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.305240 4559 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.305258 4559 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.305267 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwp2n\" (UniqueName: \"kubernetes.io/projected/8c43876e-183f-49ad-9b4d-316f1d076bb2-kube-api-access-mwp2n\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.305277 4559 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c43876e-183f-49ad-9b4d-316f1d076bb2-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.776409 4559 generic.go:334] "Generic (PLEG): container finished" podID="8c43876e-183f-49ad-9b4d-316f1d076bb2" containerID="c66de58ae6923b9aee034fc32e7b6b9a90dc9e25570679035586a68f97d2c804" exitCode=0 Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.776454 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c43876e-183f-49ad-9b4d-316f1d076bb2","Type":"ContainerDied","Data":"c66de58ae6923b9aee034fc32e7b6b9a90dc9e25570679035586a68f97d2c804"} Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.776483 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c43876e-183f-49ad-9b4d-316f1d076bb2","Type":"ContainerDied","Data":"c77bd2cd04fe1c20ccb7498adbb3151388f5a11c5ff8388f0397df2229328862"} Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.776504 4559 scope.go:117] "RemoveContainer" containerID="52784fa03ce3137011aaef3af96c67e0238d5f167260baafb042a2c088f5aa38" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.776612 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.817361 4559 scope.go:117] "RemoveContainer" containerID="6797922434b536426e5ec7f1b846712bddf13cb8f90832b7e9fcdf30de645cea" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.828884 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.835146 4559 scope.go:117] "RemoveContainer" containerID="a7392d9b52b44c2422c67927827e2f5d5cc98c6e01f107095bd6923a68020b98" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.844094 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.852042 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:01:50 crc kubenswrapper[4559]: E1123 07:01:50.852479 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c43876e-183f-49ad-9b4d-316f1d076bb2" containerName="sg-core" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.852499 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c43876e-183f-49ad-9b4d-316f1d076bb2" containerName="sg-core" Nov 23 07:01:50 crc kubenswrapper[4559]: E1123 07:01:50.852514 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c43876e-183f-49ad-9b4d-316f1d076bb2" containerName="proxy-httpd" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.852521 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c43876e-183f-49ad-9b4d-316f1d076bb2" containerName="proxy-httpd" Nov 23 07:01:50 crc kubenswrapper[4559]: E1123 07:01:50.852528 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c43876e-183f-49ad-9b4d-316f1d076bb2" containerName="ceilometer-notification-agent" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.852534 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c43876e-183f-49ad-9b4d-316f1d076bb2" containerName="ceilometer-notification-agent" Nov 23 07:01:50 crc kubenswrapper[4559]: E1123 07:01:50.852570 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c43876e-183f-49ad-9b4d-316f1d076bb2" containerName="ceilometer-central-agent" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.852576 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c43876e-183f-49ad-9b4d-316f1d076bb2" containerName="ceilometer-central-agent" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.852783 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c43876e-183f-49ad-9b4d-316f1d076bb2" containerName="ceilometer-central-agent" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.852802 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c43876e-183f-49ad-9b4d-316f1d076bb2" containerName="sg-core" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.852814 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c43876e-183f-49ad-9b4d-316f1d076bb2" containerName="proxy-httpd" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.852825 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c43876e-183f-49ad-9b4d-316f1d076bb2" containerName="ceilometer-notification-agent" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.854394 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.855931 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.858950 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.860316 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.860499 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.881456 4559 scope.go:117] "RemoveContainer" containerID="c66de58ae6923b9aee034fc32e7b6b9a90dc9e25570679035586a68f97d2c804" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.907896 4559 scope.go:117] "RemoveContainer" containerID="52784fa03ce3137011aaef3af96c67e0238d5f167260baafb042a2c088f5aa38" Nov 23 07:01:50 crc kubenswrapper[4559]: E1123 07:01:50.908534 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52784fa03ce3137011aaef3af96c67e0238d5f167260baafb042a2c088f5aa38\": container with ID starting with 52784fa03ce3137011aaef3af96c67e0238d5f167260baafb042a2c088f5aa38 not found: ID does not exist" containerID="52784fa03ce3137011aaef3af96c67e0238d5f167260baafb042a2c088f5aa38" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.908577 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52784fa03ce3137011aaef3af96c67e0238d5f167260baafb042a2c088f5aa38"} err="failed to get container status \"52784fa03ce3137011aaef3af96c67e0238d5f167260baafb042a2c088f5aa38\": rpc error: code = NotFound desc = could not find container \"52784fa03ce3137011aaef3af96c67e0238d5f167260baafb042a2c088f5aa38\": container with ID starting with 52784fa03ce3137011aaef3af96c67e0238d5f167260baafb042a2c088f5aa38 not found: ID does not exist" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.908602 4559 scope.go:117] "RemoveContainer" containerID="6797922434b536426e5ec7f1b846712bddf13cb8f90832b7e9fcdf30de645cea" Nov 23 07:01:50 crc kubenswrapper[4559]: E1123 07:01:50.909150 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6797922434b536426e5ec7f1b846712bddf13cb8f90832b7e9fcdf30de645cea\": container with ID starting with 6797922434b536426e5ec7f1b846712bddf13cb8f90832b7e9fcdf30de645cea not found: ID does not exist" containerID="6797922434b536426e5ec7f1b846712bddf13cb8f90832b7e9fcdf30de645cea" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.909191 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6797922434b536426e5ec7f1b846712bddf13cb8f90832b7e9fcdf30de645cea"} err="failed to get container status \"6797922434b536426e5ec7f1b846712bddf13cb8f90832b7e9fcdf30de645cea\": rpc error: code = NotFound desc = could not find container \"6797922434b536426e5ec7f1b846712bddf13cb8f90832b7e9fcdf30de645cea\": container with ID starting with 6797922434b536426e5ec7f1b846712bddf13cb8f90832b7e9fcdf30de645cea not found: ID does not exist" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.909216 4559 scope.go:117] "RemoveContainer" containerID="a7392d9b52b44c2422c67927827e2f5d5cc98c6e01f107095bd6923a68020b98" Nov 23 07:01:50 crc kubenswrapper[4559]: E1123 07:01:50.909956 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7392d9b52b44c2422c67927827e2f5d5cc98c6e01f107095bd6923a68020b98\": container with ID starting with a7392d9b52b44c2422c67927827e2f5d5cc98c6e01f107095bd6923a68020b98 not found: ID does not exist" containerID="a7392d9b52b44c2422c67927827e2f5d5cc98c6e01f107095bd6923a68020b98" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.910061 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7392d9b52b44c2422c67927827e2f5d5cc98c6e01f107095bd6923a68020b98"} err="failed to get container status \"a7392d9b52b44c2422c67927827e2f5d5cc98c6e01f107095bd6923a68020b98\": rpc error: code = NotFound desc = could not find container \"a7392d9b52b44c2422c67927827e2f5d5cc98c6e01f107095bd6923a68020b98\": container with ID starting with a7392d9b52b44c2422c67927827e2f5d5cc98c6e01f107095bd6923a68020b98 not found: ID does not exist" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.910143 4559 scope.go:117] "RemoveContainer" containerID="c66de58ae6923b9aee034fc32e7b6b9a90dc9e25570679035586a68f97d2c804" Nov 23 07:01:50 crc kubenswrapper[4559]: E1123 07:01:50.910573 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c66de58ae6923b9aee034fc32e7b6b9a90dc9e25570679035586a68f97d2c804\": container with ID starting with c66de58ae6923b9aee034fc32e7b6b9a90dc9e25570679035586a68f97d2c804 not found: ID does not exist" containerID="c66de58ae6923b9aee034fc32e7b6b9a90dc9e25570679035586a68f97d2c804" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.910617 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c66de58ae6923b9aee034fc32e7b6b9a90dc9e25570679035586a68f97d2c804"} err="failed to get container status \"c66de58ae6923b9aee034fc32e7b6b9a90dc9e25570679035586a68f97d2c804\": rpc error: code = NotFound desc = could not find container \"c66de58ae6923b9aee034fc32e7b6b9a90dc9e25570679035586a68f97d2c804\": container with ID starting with c66de58ae6923b9aee034fc32e7b6b9a90dc9e25570679035586a68f97d2c804 not found: ID does not exist" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.918421 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/865e8bed-9dfe-4f61-9c40-e2876af8f95b-run-httpd\") pod \"ceilometer-0\" (UID: \"865e8bed-9dfe-4f61-9c40-e2876af8f95b\") " pod="openstack/ceilometer-0" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.918481 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/865e8bed-9dfe-4f61-9c40-e2876af8f95b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"865e8bed-9dfe-4f61-9c40-e2876af8f95b\") " pod="openstack/ceilometer-0" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.918605 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pcnr\" (UniqueName: \"kubernetes.io/projected/865e8bed-9dfe-4f61-9c40-e2876af8f95b-kube-api-access-7pcnr\") pod \"ceilometer-0\" (UID: \"865e8bed-9dfe-4f61-9c40-e2876af8f95b\") " pod="openstack/ceilometer-0" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.918626 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/865e8bed-9dfe-4f61-9c40-e2876af8f95b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"865e8bed-9dfe-4f61-9c40-e2876af8f95b\") " pod="openstack/ceilometer-0" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.918701 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/865e8bed-9dfe-4f61-9c40-e2876af8f95b-log-httpd\") pod \"ceilometer-0\" (UID: \"865e8bed-9dfe-4f61-9c40-e2876af8f95b\") " pod="openstack/ceilometer-0" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.918780 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/865e8bed-9dfe-4f61-9c40-e2876af8f95b-config-data\") pod \"ceilometer-0\" (UID: \"865e8bed-9dfe-4f61-9c40-e2876af8f95b\") " pod="openstack/ceilometer-0" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.918849 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/865e8bed-9dfe-4f61-9c40-e2876af8f95b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"865e8bed-9dfe-4f61-9c40-e2876af8f95b\") " pod="openstack/ceilometer-0" Nov 23 07:01:50 crc kubenswrapper[4559]: I1123 07:01:50.918870 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/865e8bed-9dfe-4f61-9c40-e2876af8f95b-scripts\") pod \"ceilometer-0\" (UID: \"865e8bed-9dfe-4f61-9c40-e2876af8f95b\") " pod="openstack/ceilometer-0" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.021457 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/865e8bed-9dfe-4f61-9c40-e2876af8f95b-config-data\") pod \"ceilometer-0\" (UID: \"865e8bed-9dfe-4f61-9c40-e2876af8f95b\") " pod="openstack/ceilometer-0" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.021588 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/865e8bed-9dfe-4f61-9c40-e2876af8f95b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"865e8bed-9dfe-4f61-9c40-e2876af8f95b\") " pod="openstack/ceilometer-0" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.021624 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/865e8bed-9dfe-4f61-9c40-e2876af8f95b-scripts\") pod \"ceilometer-0\" (UID: \"865e8bed-9dfe-4f61-9c40-e2876af8f95b\") " pod="openstack/ceilometer-0" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.021706 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/865e8bed-9dfe-4f61-9c40-e2876af8f95b-run-httpd\") pod \"ceilometer-0\" (UID: \"865e8bed-9dfe-4f61-9c40-e2876af8f95b\") " pod="openstack/ceilometer-0" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.021759 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/865e8bed-9dfe-4f61-9c40-e2876af8f95b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"865e8bed-9dfe-4f61-9c40-e2876af8f95b\") " pod="openstack/ceilometer-0" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.021888 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pcnr\" (UniqueName: \"kubernetes.io/projected/865e8bed-9dfe-4f61-9c40-e2876af8f95b-kube-api-access-7pcnr\") pod \"ceilometer-0\" (UID: \"865e8bed-9dfe-4f61-9c40-e2876af8f95b\") " pod="openstack/ceilometer-0" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.021913 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/865e8bed-9dfe-4f61-9c40-e2876af8f95b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"865e8bed-9dfe-4f61-9c40-e2876af8f95b\") " pod="openstack/ceilometer-0" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.021969 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/865e8bed-9dfe-4f61-9c40-e2876af8f95b-log-httpd\") pod \"ceilometer-0\" (UID: \"865e8bed-9dfe-4f61-9c40-e2876af8f95b\") " pod="openstack/ceilometer-0" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.022507 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/865e8bed-9dfe-4f61-9c40-e2876af8f95b-log-httpd\") pod \"ceilometer-0\" (UID: \"865e8bed-9dfe-4f61-9c40-e2876af8f95b\") " pod="openstack/ceilometer-0" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.022783 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/865e8bed-9dfe-4f61-9c40-e2876af8f95b-run-httpd\") pod \"ceilometer-0\" (UID: \"865e8bed-9dfe-4f61-9c40-e2876af8f95b\") " pod="openstack/ceilometer-0" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.025507 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/865e8bed-9dfe-4f61-9c40-e2876af8f95b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"865e8bed-9dfe-4f61-9c40-e2876af8f95b\") " pod="openstack/ceilometer-0" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.026052 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/865e8bed-9dfe-4f61-9c40-e2876af8f95b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"865e8bed-9dfe-4f61-9c40-e2876af8f95b\") " pod="openstack/ceilometer-0" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.026283 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/865e8bed-9dfe-4f61-9c40-e2876af8f95b-config-data\") pod \"ceilometer-0\" (UID: \"865e8bed-9dfe-4f61-9c40-e2876af8f95b\") " pod="openstack/ceilometer-0" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.028826 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/865e8bed-9dfe-4f61-9c40-e2876af8f95b-scripts\") pod \"ceilometer-0\" (UID: \"865e8bed-9dfe-4f61-9c40-e2876af8f95b\") " pod="openstack/ceilometer-0" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.031426 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/865e8bed-9dfe-4f61-9c40-e2876af8f95b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"865e8bed-9dfe-4f61-9c40-e2876af8f95b\") " pod="openstack/ceilometer-0" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.038252 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pcnr\" (UniqueName: \"kubernetes.io/projected/865e8bed-9dfe-4f61-9c40-e2876af8f95b-kube-api-access-7pcnr\") pod \"ceilometer-0\" (UID: \"865e8bed-9dfe-4f61-9c40-e2876af8f95b\") " pod="openstack/ceilometer-0" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.172127 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.253330 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.331887 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b0260a5-8273-4d39-a06f-b1d02ff9696e-logs\") pod \"7b0260a5-8273-4d39-a06f-b1d02ff9696e\" (UID: \"7b0260a5-8273-4d39-a06f-b1d02ff9696e\") " Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.331965 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lv65h\" (UniqueName: \"kubernetes.io/projected/7b0260a5-8273-4d39-a06f-b1d02ff9696e-kube-api-access-lv65h\") pod \"7b0260a5-8273-4d39-a06f-b1d02ff9696e\" (UID: \"7b0260a5-8273-4d39-a06f-b1d02ff9696e\") " Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.332046 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b0260a5-8273-4d39-a06f-b1d02ff9696e-combined-ca-bundle\") pod \"7b0260a5-8273-4d39-a06f-b1d02ff9696e\" (UID: \"7b0260a5-8273-4d39-a06f-b1d02ff9696e\") " Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.332187 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b0260a5-8273-4d39-a06f-b1d02ff9696e-config-data\") pod \"7b0260a5-8273-4d39-a06f-b1d02ff9696e\" (UID: \"7b0260a5-8273-4d39-a06f-b1d02ff9696e\") " Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.337143 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b0260a5-8273-4d39-a06f-b1d02ff9696e-logs" (OuterVolumeSpecName: "logs") pod "7b0260a5-8273-4d39-a06f-b1d02ff9696e" (UID: "7b0260a5-8273-4d39-a06f-b1d02ff9696e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.340922 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b0260a5-8273-4d39-a06f-b1d02ff9696e-kube-api-access-lv65h" (OuterVolumeSpecName: "kube-api-access-lv65h") pod "7b0260a5-8273-4d39-a06f-b1d02ff9696e" (UID: "7b0260a5-8273-4d39-a06f-b1d02ff9696e"). InnerVolumeSpecName "kube-api-access-lv65h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.360860 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b0260a5-8273-4d39-a06f-b1d02ff9696e-config-data" (OuterVolumeSpecName: "config-data") pod "7b0260a5-8273-4d39-a06f-b1d02ff9696e" (UID: "7b0260a5-8273-4d39-a06f-b1d02ff9696e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.375342 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b0260a5-8273-4d39-a06f-b1d02ff9696e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7b0260a5-8273-4d39-a06f-b1d02ff9696e" (UID: "7b0260a5-8273-4d39-a06f-b1d02ff9696e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.435319 4559 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b0260a5-8273-4d39-a06f-b1d02ff9696e-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.435371 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lv65h\" (UniqueName: \"kubernetes.io/projected/7b0260a5-8273-4d39-a06f-b1d02ff9696e-kube-api-access-lv65h\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.435387 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b0260a5-8273-4d39-a06f-b1d02ff9696e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.435405 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b0260a5-8273-4d39-a06f-b1d02ff9696e-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.622052 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 07:01:51 crc kubenswrapper[4559]: W1123 07:01:51.625058 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod865e8bed_9dfe_4f61_9c40_e2876af8f95b.slice/crio-9edc935228a4cde2e9024f4e85d86eb001761e52c25cd4a69a41e50a7e0cc219 WatchSource:0}: Error finding container 9edc935228a4cde2e9024f4e85d86eb001761e52c25cd4a69a41e50a7e0cc219: Status 404 returned error can't find the container with id 9edc935228a4cde2e9024f4e85d86eb001761e52c25cd4a69a41e50a7e0cc219 Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.787860 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"865e8bed-9dfe-4f61-9c40-e2876af8f95b","Type":"ContainerStarted","Data":"9edc935228a4cde2e9024f4e85d86eb001761e52c25cd4a69a41e50a7e0cc219"} Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.790312 4559 generic.go:334] "Generic (PLEG): container finished" podID="7b0260a5-8273-4d39-a06f-b1d02ff9696e" containerID="2a5677e783c8afdf414c7343add7bd352ebead848a4aff47acdacb3119347ebe" exitCode=0 Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.790359 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7b0260a5-8273-4d39-a06f-b1d02ff9696e","Type":"ContainerDied","Data":"2a5677e783c8afdf414c7343add7bd352ebead848a4aff47acdacb3119347ebe"} Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.790365 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.790391 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7b0260a5-8273-4d39-a06f-b1d02ff9696e","Type":"ContainerDied","Data":"d13a722723e93a451712aeff55e4d7d5aa41689c6a976caedb38131bdb4231ff"} Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.790412 4559 scope.go:117] "RemoveContainer" containerID="2a5677e783c8afdf414c7343add7bd352ebead848a4aff47acdacb3119347ebe" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.831335 4559 scope.go:117] "RemoveContainer" containerID="75e3395816a677c9b178c87af51011cedca02a3b7d307f0a6cb1e864de5b6886" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.836962 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.843243 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.851785 4559 scope.go:117] "RemoveContainer" containerID="2a5677e783c8afdf414c7343add7bd352ebead848a4aff47acdacb3119347ebe" Nov 23 07:01:51 crc kubenswrapper[4559]: E1123 07:01:51.852260 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a5677e783c8afdf414c7343add7bd352ebead848a4aff47acdacb3119347ebe\": container with ID starting with 2a5677e783c8afdf414c7343add7bd352ebead848a4aff47acdacb3119347ebe not found: ID does not exist" containerID="2a5677e783c8afdf414c7343add7bd352ebead848a4aff47acdacb3119347ebe" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.852293 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a5677e783c8afdf414c7343add7bd352ebead848a4aff47acdacb3119347ebe"} err="failed to get container status \"2a5677e783c8afdf414c7343add7bd352ebead848a4aff47acdacb3119347ebe\": rpc error: code = NotFound desc = could not find container \"2a5677e783c8afdf414c7343add7bd352ebead848a4aff47acdacb3119347ebe\": container with ID starting with 2a5677e783c8afdf414c7343add7bd352ebead848a4aff47acdacb3119347ebe not found: ID does not exist" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.852313 4559 scope.go:117] "RemoveContainer" containerID="75e3395816a677c9b178c87af51011cedca02a3b7d307f0a6cb1e864de5b6886" Nov 23 07:01:51 crc kubenswrapper[4559]: E1123 07:01:51.852682 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75e3395816a677c9b178c87af51011cedca02a3b7d307f0a6cb1e864de5b6886\": container with ID starting with 75e3395816a677c9b178c87af51011cedca02a3b7d307f0a6cb1e864de5b6886 not found: ID does not exist" containerID="75e3395816a677c9b178c87af51011cedca02a3b7d307f0a6cb1e864de5b6886" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.852714 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75e3395816a677c9b178c87af51011cedca02a3b7d307f0a6cb1e864de5b6886"} err="failed to get container status \"75e3395816a677c9b178c87af51011cedca02a3b7d307f0a6cb1e864de5b6886\": rpc error: code = NotFound desc = could not find container \"75e3395816a677c9b178c87af51011cedca02a3b7d307f0a6cb1e864de5b6886\": container with ID starting with 75e3395816a677c9b178c87af51011cedca02a3b7d307f0a6cb1e864de5b6886 not found: ID does not exist" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.857503 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 23 07:01:51 crc kubenswrapper[4559]: E1123 07:01:51.857891 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b0260a5-8273-4d39-a06f-b1d02ff9696e" containerName="nova-api-api" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.857909 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b0260a5-8273-4d39-a06f-b1d02ff9696e" containerName="nova-api-api" Nov 23 07:01:51 crc kubenswrapper[4559]: E1123 07:01:51.857925 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b0260a5-8273-4d39-a06f-b1d02ff9696e" containerName="nova-api-log" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.857932 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b0260a5-8273-4d39-a06f-b1d02ff9696e" containerName="nova-api-log" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.858106 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b0260a5-8273-4d39-a06f-b1d02ff9696e" containerName="nova-api-api" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.858123 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b0260a5-8273-4d39-a06f-b1d02ff9696e" containerName="nova-api-log" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.859003 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.866095 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.866627 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.866850 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.866977 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.949210 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a4a109b-31a8-42a0-a025-05652e26b669-public-tls-certs\") pod \"nova-api-0\" (UID: \"8a4a109b-31a8-42a0-a025-05652e26b669\") " pod="openstack/nova-api-0" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.949264 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a4a109b-31a8-42a0-a025-05652e26b669-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8a4a109b-31a8-42a0-a025-05652e26b669\") " pod="openstack/nova-api-0" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.949284 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cg9s\" (UniqueName: \"kubernetes.io/projected/8a4a109b-31a8-42a0-a025-05652e26b669-kube-api-access-7cg9s\") pod \"nova-api-0\" (UID: \"8a4a109b-31a8-42a0-a025-05652e26b669\") " pod="openstack/nova-api-0" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.949309 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a4a109b-31a8-42a0-a025-05652e26b669-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8a4a109b-31a8-42a0-a025-05652e26b669\") " pod="openstack/nova-api-0" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.949373 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a4a109b-31a8-42a0-a025-05652e26b669-logs\") pod \"nova-api-0\" (UID: \"8a4a109b-31a8-42a0-a025-05652e26b669\") " pod="openstack/nova-api-0" Nov 23 07:01:51 crc kubenswrapper[4559]: I1123 07:01:51.949391 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a4a109b-31a8-42a0-a025-05652e26b669-config-data\") pod \"nova-api-0\" (UID: \"8a4a109b-31a8-42a0-a025-05652e26b669\") " pod="openstack/nova-api-0" Nov 23 07:01:52 crc kubenswrapper[4559]: I1123 07:01:52.050781 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a4a109b-31a8-42a0-a025-05652e26b669-public-tls-certs\") pod \"nova-api-0\" (UID: \"8a4a109b-31a8-42a0-a025-05652e26b669\") " pod="openstack/nova-api-0" Nov 23 07:01:52 crc kubenswrapper[4559]: I1123 07:01:52.050850 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a4a109b-31a8-42a0-a025-05652e26b669-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8a4a109b-31a8-42a0-a025-05652e26b669\") " pod="openstack/nova-api-0" Nov 23 07:01:52 crc kubenswrapper[4559]: I1123 07:01:52.050871 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cg9s\" (UniqueName: \"kubernetes.io/projected/8a4a109b-31a8-42a0-a025-05652e26b669-kube-api-access-7cg9s\") pod \"nova-api-0\" (UID: \"8a4a109b-31a8-42a0-a025-05652e26b669\") " pod="openstack/nova-api-0" Nov 23 07:01:52 crc kubenswrapper[4559]: I1123 07:01:52.050900 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a4a109b-31a8-42a0-a025-05652e26b669-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8a4a109b-31a8-42a0-a025-05652e26b669\") " pod="openstack/nova-api-0" Nov 23 07:01:52 crc kubenswrapper[4559]: I1123 07:01:52.051019 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a4a109b-31a8-42a0-a025-05652e26b669-logs\") pod \"nova-api-0\" (UID: \"8a4a109b-31a8-42a0-a025-05652e26b669\") " pod="openstack/nova-api-0" Nov 23 07:01:52 crc kubenswrapper[4559]: I1123 07:01:52.051045 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a4a109b-31a8-42a0-a025-05652e26b669-config-data\") pod \"nova-api-0\" (UID: \"8a4a109b-31a8-42a0-a025-05652e26b669\") " pod="openstack/nova-api-0" Nov 23 07:01:52 crc kubenswrapper[4559]: I1123 07:01:52.051564 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a4a109b-31a8-42a0-a025-05652e26b669-logs\") pod \"nova-api-0\" (UID: \"8a4a109b-31a8-42a0-a025-05652e26b669\") " pod="openstack/nova-api-0" Nov 23 07:01:52 crc kubenswrapper[4559]: I1123 07:01:52.054808 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a4a109b-31a8-42a0-a025-05652e26b669-public-tls-certs\") pod \"nova-api-0\" (UID: \"8a4a109b-31a8-42a0-a025-05652e26b669\") " pod="openstack/nova-api-0" Nov 23 07:01:52 crc kubenswrapper[4559]: I1123 07:01:52.055364 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a4a109b-31a8-42a0-a025-05652e26b669-config-data\") pod \"nova-api-0\" (UID: \"8a4a109b-31a8-42a0-a025-05652e26b669\") " pod="openstack/nova-api-0" Nov 23 07:01:52 crc kubenswrapper[4559]: I1123 07:01:52.055673 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a4a109b-31a8-42a0-a025-05652e26b669-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8a4a109b-31a8-42a0-a025-05652e26b669\") " pod="openstack/nova-api-0" Nov 23 07:01:52 crc kubenswrapper[4559]: I1123 07:01:52.055770 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a4a109b-31a8-42a0-a025-05652e26b669-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8a4a109b-31a8-42a0-a025-05652e26b669\") " pod="openstack/nova-api-0" Nov 23 07:01:52 crc kubenswrapper[4559]: I1123 07:01:52.069923 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cg9s\" (UniqueName: \"kubernetes.io/projected/8a4a109b-31a8-42a0-a025-05652e26b669-kube-api-access-7cg9s\") pod \"nova-api-0\" (UID: \"8a4a109b-31a8-42a0-a025-05652e26b669\") " pod="openstack/nova-api-0" Nov 23 07:01:52 crc kubenswrapper[4559]: I1123 07:01:52.172808 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:01:52 crc kubenswrapper[4559]: I1123 07:01:52.285444 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b0260a5-8273-4d39-a06f-b1d02ff9696e" path="/var/lib/kubelet/pods/7b0260a5-8273-4d39-a06f-b1d02ff9696e/volumes" Nov 23 07:01:52 crc kubenswrapper[4559]: I1123 07:01:52.286545 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c43876e-183f-49ad-9b4d-316f1d076bb2" path="/var/lib/kubelet/pods/8c43876e-183f-49ad-9b4d-316f1d076bb2/volumes" Nov 23 07:01:52 crc kubenswrapper[4559]: I1123 07:01:52.405363 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:52 crc kubenswrapper[4559]: I1123 07:01:52.615070 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:01:52 crc kubenswrapper[4559]: W1123 07:01:52.618958 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a4a109b_31a8_42a0_a025_05652e26b669.slice/crio-04713f5599ce16ae416fc50fe4f76e93c329e50fad646746e4a80037acbf6dde WatchSource:0}: Error finding container 04713f5599ce16ae416fc50fe4f76e93c329e50fad646746e4a80037acbf6dde: Status 404 returned error can't find the container with id 04713f5599ce16ae416fc50fe4f76e93c329e50fad646746e4a80037acbf6dde Nov 23 07:01:52 crc kubenswrapper[4559]: I1123 07:01:52.804338 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8a4a109b-31a8-42a0-a025-05652e26b669","Type":"ContainerStarted","Data":"b0fdf7c9abb4776f1a0d53e9dc3458a5f19f90608d1edf65d82f2b3aa8bb6185"} Nov 23 07:01:52 crc kubenswrapper[4559]: I1123 07:01:52.804381 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8a4a109b-31a8-42a0-a025-05652e26b669","Type":"ContainerStarted","Data":"04713f5599ce16ae416fc50fe4f76e93c329e50fad646746e4a80037acbf6dde"} Nov 23 07:01:52 crc kubenswrapper[4559]: I1123 07:01:52.806093 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"865e8bed-9dfe-4f61-9c40-e2876af8f95b","Type":"ContainerStarted","Data":"6ecd74c3b825efbb39eef9d41c6abc31080f3428f3158ec2713c7d08376c39c9"} Nov 23 07:01:53 crc kubenswrapper[4559]: I1123 07:01:53.816432 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"865e8bed-9dfe-4f61-9c40-e2876af8f95b","Type":"ContainerStarted","Data":"4af9dfcb22a3f0cc0520e8851744d21842f30a44f9161dd8801aab37bf037180"} Nov 23 07:01:53 crc kubenswrapper[4559]: I1123 07:01:53.818264 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8a4a109b-31a8-42a0-a025-05652e26b669","Type":"ContainerStarted","Data":"2b4ac7acd62790e8aea151ad93fa0fe2bfc3459660cc3f7946b8911783aa556d"} Nov 23 07:01:53 crc kubenswrapper[4559]: I1123 07:01:53.841994 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.841971146 podStartE2EDuration="2.841971146s" podCreationTimestamp="2025-11-23 07:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:01:53.839743367 +0000 UTC m=+1015.861728981" watchObservedRunningTime="2025-11-23 07:01:53.841971146 +0000 UTC m=+1015.863956761" Nov 23 07:01:54 crc kubenswrapper[4559]: I1123 07:01:54.831904 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"865e8bed-9dfe-4f61-9c40-e2876af8f95b","Type":"ContainerStarted","Data":"48fd9d73b8f34572abe361c998797db9dc30f6b51d3d85be0494cc319ede8b83"} Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.189822 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.251819 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77cbccc799-6bxp4"] Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.252154 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" podUID="9e626278-ea64-4839-8cf2-852c7d336dbb" containerName="dnsmasq-dns" containerID="cri-o://0deb60e939ea52ab20cf717534bd8464a2f2a8a8bde2111d095b2e77ce4b3cef" gracePeriod=10 Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.707730 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.836602 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8n6zn\" (UniqueName: \"kubernetes.io/projected/9e626278-ea64-4839-8cf2-852c7d336dbb-kube-api-access-8n6zn\") pod \"9e626278-ea64-4839-8cf2-852c7d336dbb\" (UID: \"9e626278-ea64-4839-8cf2-852c7d336dbb\") " Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.836720 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-dns-svc\") pod \"9e626278-ea64-4839-8cf2-852c7d336dbb\" (UID: \"9e626278-ea64-4839-8cf2-852c7d336dbb\") " Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.836743 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-ovsdbserver-sb\") pod \"9e626278-ea64-4839-8cf2-852c7d336dbb\" (UID: \"9e626278-ea64-4839-8cf2-852c7d336dbb\") " Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.836816 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-dns-swift-storage-0\") pod \"9e626278-ea64-4839-8cf2-852c7d336dbb\" (UID: \"9e626278-ea64-4839-8cf2-852c7d336dbb\") " Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.836896 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-ovsdbserver-nb\") pod \"9e626278-ea64-4839-8cf2-852c7d336dbb\" (UID: \"9e626278-ea64-4839-8cf2-852c7d336dbb\") " Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.837061 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-config\") pod \"9e626278-ea64-4839-8cf2-852c7d336dbb\" (UID: \"9e626278-ea64-4839-8cf2-852c7d336dbb\") " Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.858831 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e626278-ea64-4839-8cf2-852c7d336dbb-kube-api-access-8n6zn" (OuterVolumeSpecName: "kube-api-access-8n6zn") pod "9e626278-ea64-4839-8cf2-852c7d336dbb" (UID: "9e626278-ea64-4839-8cf2-852c7d336dbb"). InnerVolumeSpecName "kube-api-access-8n6zn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.868673 4559 generic.go:334] "Generic (PLEG): container finished" podID="9e626278-ea64-4839-8cf2-852c7d336dbb" containerID="0deb60e939ea52ab20cf717534bd8464a2f2a8a8bde2111d095b2e77ce4b3cef" exitCode=0 Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.868761 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" event={"ID":"9e626278-ea64-4839-8cf2-852c7d336dbb","Type":"ContainerDied","Data":"0deb60e939ea52ab20cf717534bd8464a2f2a8a8bde2111d095b2e77ce4b3cef"} Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.868792 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" event={"ID":"9e626278-ea64-4839-8cf2-852c7d336dbb","Type":"ContainerDied","Data":"e0fbc0d60bd28f91402efdd653cf42c3c74f5270cb28b278880cc937e4fdf216"} Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.868810 4559 scope.go:117] "RemoveContainer" containerID="0deb60e939ea52ab20cf717534bd8464a2f2a8a8bde2111d095b2e77ce4b3cef" Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.868932 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77cbccc799-6bxp4" Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.913785 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"865e8bed-9dfe-4f61-9c40-e2876af8f95b","Type":"ContainerStarted","Data":"d7588b7b130db6fff7f115fd3b3d9711bac261b8f77f06858f34423903fd1c7e"} Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.915299 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.927771 4559 scope.go:117] "RemoveContainer" containerID="558e702dfe65ebde1c55086eb1be211da8e0357ffe2640812361dc238e9ce254" Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.938119 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9e626278-ea64-4839-8cf2-852c7d336dbb" (UID: "9e626278-ea64-4839-8cf2-852c7d336dbb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.938334 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-dns-swift-storage-0\") pod \"9e626278-ea64-4839-8cf2-852c7d336dbb\" (UID: \"9e626278-ea64-4839-8cf2-852c7d336dbb\") " Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.939065 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8n6zn\" (UniqueName: \"kubernetes.io/projected/9e626278-ea64-4839-8cf2-852c7d336dbb-kube-api-access-8n6zn\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:55 crc kubenswrapper[4559]: W1123 07:01:55.939164 4559 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/9e626278-ea64-4839-8cf2-852c7d336dbb/volumes/kubernetes.io~configmap/dns-swift-storage-0 Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.939184 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9e626278-ea64-4839-8cf2-852c7d336dbb" (UID: "9e626278-ea64-4839-8cf2-852c7d336dbb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.943821 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.344714174 podStartE2EDuration="5.943802595s" podCreationTimestamp="2025-11-23 07:01:50 +0000 UTC" firstStartedPulling="2025-11-23 07:01:51.628516356 +0000 UTC m=+1013.650501971" lastFinishedPulling="2025-11-23 07:01:55.227604779 +0000 UTC m=+1017.249590392" observedRunningTime="2025-11-23 07:01:55.940200931 +0000 UTC m=+1017.962186545" watchObservedRunningTime="2025-11-23 07:01:55.943802595 +0000 UTC m=+1017.965788208" Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.971312 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9e626278-ea64-4839-8cf2-852c7d336dbb" (UID: "9e626278-ea64-4839-8cf2-852c7d336dbb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.984055 4559 scope.go:117] "RemoveContainer" containerID="0deb60e939ea52ab20cf717534bd8464a2f2a8a8bde2111d095b2e77ce4b3cef" Nov 23 07:01:55 crc kubenswrapper[4559]: E1123 07:01:55.984544 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0deb60e939ea52ab20cf717534bd8464a2f2a8a8bde2111d095b2e77ce4b3cef\": container with ID starting with 0deb60e939ea52ab20cf717534bd8464a2f2a8a8bde2111d095b2e77ce4b3cef not found: ID does not exist" containerID="0deb60e939ea52ab20cf717534bd8464a2f2a8a8bde2111d095b2e77ce4b3cef" Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.984589 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0deb60e939ea52ab20cf717534bd8464a2f2a8a8bde2111d095b2e77ce4b3cef"} err="failed to get container status \"0deb60e939ea52ab20cf717534bd8464a2f2a8a8bde2111d095b2e77ce4b3cef\": rpc error: code = NotFound desc = could not find container \"0deb60e939ea52ab20cf717534bd8464a2f2a8a8bde2111d095b2e77ce4b3cef\": container with ID starting with 0deb60e939ea52ab20cf717534bd8464a2f2a8a8bde2111d095b2e77ce4b3cef not found: ID does not exist" Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.984618 4559 scope.go:117] "RemoveContainer" containerID="558e702dfe65ebde1c55086eb1be211da8e0357ffe2640812361dc238e9ce254" Nov 23 07:01:55 crc kubenswrapper[4559]: E1123 07:01:55.985072 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"558e702dfe65ebde1c55086eb1be211da8e0357ffe2640812361dc238e9ce254\": container with ID starting with 558e702dfe65ebde1c55086eb1be211da8e0357ffe2640812361dc238e9ce254 not found: ID does not exist" containerID="558e702dfe65ebde1c55086eb1be211da8e0357ffe2640812361dc238e9ce254" Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.985116 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"558e702dfe65ebde1c55086eb1be211da8e0357ffe2640812361dc238e9ce254"} err="failed to get container status \"558e702dfe65ebde1c55086eb1be211da8e0357ffe2640812361dc238e9ce254\": rpc error: code = NotFound desc = could not find container \"558e702dfe65ebde1c55086eb1be211da8e0357ffe2640812361dc238e9ce254\": container with ID starting with 558e702dfe65ebde1c55086eb1be211da8e0357ffe2640812361dc238e9ce254 not found: ID does not exist" Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.985552 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-config" (OuterVolumeSpecName: "config") pod "9e626278-ea64-4839-8cf2-852c7d336dbb" (UID: "9e626278-ea64-4839-8cf2-852c7d336dbb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:01:55 crc kubenswrapper[4559]: I1123 07:01:55.989718 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9e626278-ea64-4839-8cf2-852c7d336dbb" (UID: "9e626278-ea64-4839-8cf2-852c7d336dbb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:01:56 crc kubenswrapper[4559]: I1123 07:01:56.007165 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9e626278-ea64-4839-8cf2-852c7d336dbb" (UID: "9e626278-ea64-4839-8cf2-852c7d336dbb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:01:56 crc kubenswrapper[4559]: I1123 07:01:56.042361 4559 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:56 crc kubenswrapper[4559]: I1123 07:01:56.042394 4559 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:56 crc kubenswrapper[4559]: I1123 07:01:56.042412 4559 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:56 crc kubenswrapper[4559]: I1123 07:01:56.042425 4559 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:56 crc kubenswrapper[4559]: I1123 07:01:56.042436 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e626278-ea64-4839-8cf2-852c7d336dbb-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:01:56 crc kubenswrapper[4559]: I1123 07:01:56.197200 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77cbccc799-6bxp4"] Nov 23 07:01:56 crc kubenswrapper[4559]: I1123 07:01:56.202467 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77cbccc799-6bxp4"] Nov 23 07:01:56 crc kubenswrapper[4559]: I1123 07:01:56.284137 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e626278-ea64-4839-8cf2-852c7d336dbb" path="/var/lib/kubelet/pods/9e626278-ea64-4839-8cf2-852c7d336dbb/volumes" Nov 23 07:01:57 crc kubenswrapper[4559]: I1123 07:01:57.405746 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:57 crc kubenswrapper[4559]: I1123 07:01:57.429874 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:57 crc kubenswrapper[4559]: I1123 07:01:57.948558 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 23 07:01:58 crc kubenswrapper[4559]: I1123 07:01:58.070826 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-msr7f"] Nov 23 07:01:58 crc kubenswrapper[4559]: E1123 07:01:58.071339 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e626278-ea64-4839-8cf2-852c7d336dbb" containerName="init" Nov 23 07:01:58 crc kubenswrapper[4559]: I1123 07:01:58.071361 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e626278-ea64-4839-8cf2-852c7d336dbb" containerName="init" Nov 23 07:01:58 crc kubenswrapper[4559]: E1123 07:01:58.071402 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e626278-ea64-4839-8cf2-852c7d336dbb" containerName="dnsmasq-dns" Nov 23 07:01:58 crc kubenswrapper[4559]: I1123 07:01:58.071408 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e626278-ea64-4839-8cf2-852c7d336dbb" containerName="dnsmasq-dns" Nov 23 07:01:58 crc kubenswrapper[4559]: I1123 07:01:58.071588 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e626278-ea64-4839-8cf2-852c7d336dbb" containerName="dnsmasq-dns" Nov 23 07:01:58 crc kubenswrapper[4559]: I1123 07:01:58.072268 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-msr7f" Nov 23 07:01:58 crc kubenswrapper[4559]: I1123 07:01:58.075381 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 23 07:01:58 crc kubenswrapper[4559]: I1123 07:01:58.075569 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 23 07:01:58 crc kubenswrapper[4559]: I1123 07:01:58.076930 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-msr7f"] Nov 23 07:01:58 crc kubenswrapper[4559]: I1123 07:01:58.186708 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jhbt\" (UniqueName: \"kubernetes.io/projected/c818ed2a-d851-4a13-b67d-967f178654dc-kube-api-access-6jhbt\") pod \"nova-cell1-cell-mapping-msr7f\" (UID: \"c818ed2a-d851-4a13-b67d-967f178654dc\") " pod="openstack/nova-cell1-cell-mapping-msr7f" Nov 23 07:01:58 crc kubenswrapper[4559]: I1123 07:01:58.186789 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c818ed2a-d851-4a13-b67d-967f178654dc-scripts\") pod \"nova-cell1-cell-mapping-msr7f\" (UID: \"c818ed2a-d851-4a13-b67d-967f178654dc\") " pod="openstack/nova-cell1-cell-mapping-msr7f" Nov 23 07:01:58 crc kubenswrapper[4559]: I1123 07:01:58.186898 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c818ed2a-d851-4a13-b67d-967f178654dc-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-msr7f\" (UID: \"c818ed2a-d851-4a13-b67d-967f178654dc\") " pod="openstack/nova-cell1-cell-mapping-msr7f" Nov 23 07:01:58 crc kubenswrapper[4559]: I1123 07:01:58.187001 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c818ed2a-d851-4a13-b67d-967f178654dc-config-data\") pod \"nova-cell1-cell-mapping-msr7f\" (UID: \"c818ed2a-d851-4a13-b67d-967f178654dc\") " pod="openstack/nova-cell1-cell-mapping-msr7f" Nov 23 07:01:58 crc kubenswrapper[4559]: I1123 07:01:58.289667 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jhbt\" (UniqueName: \"kubernetes.io/projected/c818ed2a-d851-4a13-b67d-967f178654dc-kube-api-access-6jhbt\") pod \"nova-cell1-cell-mapping-msr7f\" (UID: \"c818ed2a-d851-4a13-b67d-967f178654dc\") " pod="openstack/nova-cell1-cell-mapping-msr7f" Nov 23 07:01:58 crc kubenswrapper[4559]: I1123 07:01:58.289772 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c818ed2a-d851-4a13-b67d-967f178654dc-scripts\") pod \"nova-cell1-cell-mapping-msr7f\" (UID: \"c818ed2a-d851-4a13-b67d-967f178654dc\") " pod="openstack/nova-cell1-cell-mapping-msr7f" Nov 23 07:01:58 crc kubenswrapper[4559]: I1123 07:01:58.289820 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c818ed2a-d851-4a13-b67d-967f178654dc-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-msr7f\" (UID: \"c818ed2a-d851-4a13-b67d-967f178654dc\") " pod="openstack/nova-cell1-cell-mapping-msr7f" Nov 23 07:01:58 crc kubenswrapper[4559]: I1123 07:01:58.289862 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c818ed2a-d851-4a13-b67d-967f178654dc-config-data\") pod \"nova-cell1-cell-mapping-msr7f\" (UID: \"c818ed2a-d851-4a13-b67d-967f178654dc\") " pod="openstack/nova-cell1-cell-mapping-msr7f" Nov 23 07:01:58 crc kubenswrapper[4559]: I1123 07:01:58.292288 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 23 07:01:58 crc kubenswrapper[4559]: I1123 07:01:58.292772 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 23 07:01:58 crc kubenswrapper[4559]: I1123 07:01:58.301845 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c818ed2a-d851-4a13-b67d-967f178654dc-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-msr7f\" (UID: \"c818ed2a-d851-4a13-b67d-967f178654dc\") " pod="openstack/nova-cell1-cell-mapping-msr7f" Nov 23 07:01:58 crc kubenswrapper[4559]: I1123 07:01:58.303873 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c818ed2a-d851-4a13-b67d-967f178654dc-scripts\") pod \"nova-cell1-cell-mapping-msr7f\" (UID: \"c818ed2a-d851-4a13-b67d-967f178654dc\") " pod="openstack/nova-cell1-cell-mapping-msr7f" Nov 23 07:01:58 crc kubenswrapper[4559]: I1123 07:01:58.305972 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c818ed2a-d851-4a13-b67d-967f178654dc-config-data\") pod \"nova-cell1-cell-mapping-msr7f\" (UID: \"c818ed2a-d851-4a13-b67d-967f178654dc\") " pod="openstack/nova-cell1-cell-mapping-msr7f" Nov 23 07:01:58 crc kubenswrapper[4559]: I1123 07:01:58.306496 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jhbt\" (UniqueName: \"kubernetes.io/projected/c818ed2a-d851-4a13-b67d-967f178654dc-kube-api-access-6jhbt\") pod \"nova-cell1-cell-mapping-msr7f\" (UID: \"c818ed2a-d851-4a13-b67d-967f178654dc\") " pod="openstack/nova-cell1-cell-mapping-msr7f" Nov 23 07:01:58 crc kubenswrapper[4559]: I1123 07:01:58.393346 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-msr7f" Nov 23 07:01:58 crc kubenswrapper[4559]: I1123 07:01:58.808386 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-msr7f"] Nov 23 07:01:58 crc kubenswrapper[4559]: I1123 07:01:58.945962 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-msr7f" event={"ID":"c818ed2a-d851-4a13-b67d-967f178654dc","Type":"ContainerStarted","Data":"d3a00a4e5bddd968e8694df14367d55824a77e055222085cb272bbcc67dc5f98"} Nov 23 07:01:59 crc kubenswrapper[4559]: I1123 07:01:59.956791 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-msr7f" event={"ID":"c818ed2a-d851-4a13-b67d-967f178654dc","Type":"ContainerStarted","Data":"56edeb6a2681718e3f9d3ecb52b31a83a292944b5f25164edc1255ba37888f0b"} Nov 23 07:01:59 crc kubenswrapper[4559]: I1123 07:01:59.975846 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-msr7f" podStartSLOduration=1.975830794 podStartE2EDuration="1.975830794s" podCreationTimestamp="2025-11-23 07:01:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:01:59.969241111 +0000 UTC m=+1021.991226725" watchObservedRunningTime="2025-11-23 07:01:59.975830794 +0000 UTC m=+1021.997816407" Nov 23 07:02:02 crc kubenswrapper[4559]: I1123 07:02:02.173065 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 23 07:02:02 crc kubenswrapper[4559]: I1123 07:02:02.173317 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 23 07:02:02 crc kubenswrapper[4559]: I1123 07:02:02.990452 4559 generic.go:334] "Generic (PLEG): container finished" podID="c818ed2a-d851-4a13-b67d-967f178654dc" containerID="56edeb6a2681718e3f9d3ecb52b31a83a292944b5f25164edc1255ba37888f0b" exitCode=0 Nov 23 07:02:02 crc kubenswrapper[4559]: I1123 07:02:02.990506 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-msr7f" event={"ID":"c818ed2a-d851-4a13-b67d-967f178654dc","Type":"ContainerDied","Data":"56edeb6a2681718e3f9d3ecb52b31a83a292944b5f25164edc1255ba37888f0b"} Nov 23 07:02:03 crc kubenswrapper[4559]: I1123 07:02:03.186754 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8a4a109b-31a8-42a0-a025-05652e26b669" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.199:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 23 07:02:03 crc kubenswrapper[4559]: I1123 07:02:03.186754 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8a4a109b-31a8-42a0-a025-05652e26b669" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.199:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 23 07:02:04 crc kubenswrapper[4559]: I1123 07:02:04.293900 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-msr7f" Nov 23 07:02:04 crc kubenswrapper[4559]: I1123 07:02:04.407368 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c818ed2a-d851-4a13-b67d-967f178654dc-combined-ca-bundle\") pod \"c818ed2a-d851-4a13-b67d-967f178654dc\" (UID: \"c818ed2a-d851-4a13-b67d-967f178654dc\") " Nov 23 07:02:04 crc kubenswrapper[4559]: I1123 07:02:04.407457 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jhbt\" (UniqueName: \"kubernetes.io/projected/c818ed2a-d851-4a13-b67d-967f178654dc-kube-api-access-6jhbt\") pod \"c818ed2a-d851-4a13-b67d-967f178654dc\" (UID: \"c818ed2a-d851-4a13-b67d-967f178654dc\") " Nov 23 07:02:04 crc kubenswrapper[4559]: I1123 07:02:04.407586 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c818ed2a-d851-4a13-b67d-967f178654dc-scripts\") pod \"c818ed2a-d851-4a13-b67d-967f178654dc\" (UID: \"c818ed2a-d851-4a13-b67d-967f178654dc\") " Nov 23 07:02:04 crc kubenswrapper[4559]: I1123 07:02:04.407820 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c818ed2a-d851-4a13-b67d-967f178654dc-config-data\") pod \"c818ed2a-d851-4a13-b67d-967f178654dc\" (UID: \"c818ed2a-d851-4a13-b67d-967f178654dc\") " Nov 23 07:02:04 crc kubenswrapper[4559]: I1123 07:02:04.414084 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c818ed2a-d851-4a13-b67d-967f178654dc-scripts" (OuterVolumeSpecName: "scripts") pod "c818ed2a-d851-4a13-b67d-967f178654dc" (UID: "c818ed2a-d851-4a13-b67d-967f178654dc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:02:04 crc kubenswrapper[4559]: I1123 07:02:04.414258 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c818ed2a-d851-4a13-b67d-967f178654dc-kube-api-access-6jhbt" (OuterVolumeSpecName: "kube-api-access-6jhbt") pod "c818ed2a-d851-4a13-b67d-967f178654dc" (UID: "c818ed2a-d851-4a13-b67d-967f178654dc"). InnerVolumeSpecName "kube-api-access-6jhbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:02:04 crc kubenswrapper[4559]: I1123 07:02:04.438064 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c818ed2a-d851-4a13-b67d-967f178654dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c818ed2a-d851-4a13-b67d-967f178654dc" (UID: "c818ed2a-d851-4a13-b67d-967f178654dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:02:04 crc kubenswrapper[4559]: I1123 07:02:04.439797 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c818ed2a-d851-4a13-b67d-967f178654dc-config-data" (OuterVolumeSpecName: "config-data") pod "c818ed2a-d851-4a13-b67d-967f178654dc" (UID: "c818ed2a-d851-4a13-b67d-967f178654dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:02:04 crc kubenswrapper[4559]: I1123 07:02:04.511324 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c818ed2a-d851-4a13-b67d-967f178654dc-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:04 crc kubenswrapper[4559]: I1123 07:02:04.511363 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c818ed2a-d851-4a13-b67d-967f178654dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:04 crc kubenswrapper[4559]: I1123 07:02:04.511377 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jhbt\" (UniqueName: \"kubernetes.io/projected/c818ed2a-d851-4a13-b67d-967f178654dc-kube-api-access-6jhbt\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:04 crc kubenswrapper[4559]: I1123 07:02:04.511389 4559 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c818ed2a-d851-4a13-b67d-967f178654dc-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:05 crc kubenswrapper[4559]: I1123 07:02:05.012200 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-msr7f" event={"ID":"c818ed2a-d851-4a13-b67d-967f178654dc","Type":"ContainerDied","Data":"d3a00a4e5bddd968e8694df14367d55824a77e055222085cb272bbcc67dc5f98"} Nov 23 07:02:05 crc kubenswrapper[4559]: I1123 07:02:05.012495 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3a00a4e5bddd968e8694df14367d55824a77e055222085cb272bbcc67dc5f98" Nov 23 07:02:05 crc kubenswrapper[4559]: I1123 07:02:05.012326 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-msr7f" Nov 23 07:02:05 crc kubenswrapper[4559]: I1123 07:02:05.135942 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:02:05 crc kubenswrapper[4559]: I1123 07:02:05.136234 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="27bbb5a9-7e24-4529-a05e-1a0bdc5c7572" containerName="nova-scheduler-scheduler" containerID="cri-o://6eff3da2d979d6fd24ecae38d241fea26cc86e4a77ec0a6676a5d6e92fdc15a4" gracePeriod=30 Nov 23 07:02:05 crc kubenswrapper[4559]: I1123 07:02:05.144077 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:02:05 crc kubenswrapper[4559]: I1123 07:02:05.144356 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8a4a109b-31a8-42a0-a025-05652e26b669" containerName="nova-api-log" containerID="cri-o://b0fdf7c9abb4776f1a0d53e9dc3458a5f19f90608d1edf65d82f2b3aa8bb6185" gracePeriod=30 Nov 23 07:02:05 crc kubenswrapper[4559]: I1123 07:02:05.144538 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8a4a109b-31a8-42a0-a025-05652e26b669" containerName="nova-api-api" containerID="cri-o://2b4ac7acd62790e8aea151ad93fa0fe2bfc3459660cc3f7946b8911783aa556d" gracePeriod=30 Nov 23 07:02:05 crc kubenswrapper[4559]: I1123 07:02:05.177126 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:02:05 crc kubenswrapper[4559]: I1123 07:02:05.177491 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8d0d4e86-3876-4666-b628-83cb6c08700b" containerName="nova-metadata-log" containerID="cri-o://022d9852ba23da5dec14c28273df4974709ac8978ff19f432268af796578b513" gracePeriod=30 Nov 23 07:02:05 crc kubenswrapper[4559]: I1123 07:02:05.177617 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8d0d4e86-3876-4666-b628-83cb6c08700b" containerName="nova-metadata-metadata" containerID="cri-o://9b5836b9d27c68a11951faf7967525bf1cd78e4a672b484c343e84c3363bac02" gracePeriod=30 Nov 23 07:02:06 crc kubenswrapper[4559]: I1123 07:02:06.023549 4559 generic.go:334] "Generic (PLEG): container finished" podID="27bbb5a9-7e24-4529-a05e-1a0bdc5c7572" containerID="6eff3da2d979d6fd24ecae38d241fea26cc86e4a77ec0a6676a5d6e92fdc15a4" exitCode=0 Nov 23 07:02:06 crc kubenswrapper[4559]: I1123 07:02:06.023654 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"27bbb5a9-7e24-4529-a05e-1a0bdc5c7572","Type":"ContainerDied","Data":"6eff3da2d979d6fd24ecae38d241fea26cc86e4a77ec0a6676a5d6e92fdc15a4"} Nov 23 07:02:06 crc kubenswrapper[4559]: I1123 07:02:06.026395 4559 generic.go:334] "Generic (PLEG): container finished" podID="8a4a109b-31a8-42a0-a025-05652e26b669" containerID="b0fdf7c9abb4776f1a0d53e9dc3458a5f19f90608d1edf65d82f2b3aa8bb6185" exitCode=143 Nov 23 07:02:06 crc kubenswrapper[4559]: I1123 07:02:06.026456 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8a4a109b-31a8-42a0-a025-05652e26b669","Type":"ContainerDied","Data":"b0fdf7c9abb4776f1a0d53e9dc3458a5f19f90608d1edf65d82f2b3aa8bb6185"} Nov 23 07:02:06 crc kubenswrapper[4559]: I1123 07:02:06.028374 4559 generic.go:334] "Generic (PLEG): container finished" podID="8d0d4e86-3876-4666-b628-83cb6c08700b" containerID="022d9852ba23da5dec14c28273df4974709ac8978ff19f432268af796578b513" exitCode=143 Nov 23 07:02:06 crc kubenswrapper[4559]: I1123 07:02:06.028408 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8d0d4e86-3876-4666-b628-83cb6c08700b","Type":"ContainerDied","Data":"022d9852ba23da5dec14c28273df4974709ac8978ff19f432268af796578b513"} Nov 23 07:02:06 crc kubenswrapper[4559]: I1123 07:02:06.080330 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 07:02:06 crc kubenswrapper[4559]: I1123 07:02:06.252942 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27bbb5a9-7e24-4529-a05e-1a0bdc5c7572-combined-ca-bundle\") pod \"27bbb5a9-7e24-4529-a05e-1a0bdc5c7572\" (UID: \"27bbb5a9-7e24-4529-a05e-1a0bdc5c7572\") " Nov 23 07:02:06 crc kubenswrapper[4559]: I1123 07:02:06.253463 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27bbb5a9-7e24-4529-a05e-1a0bdc5c7572-config-data\") pod \"27bbb5a9-7e24-4529-a05e-1a0bdc5c7572\" (UID: \"27bbb5a9-7e24-4529-a05e-1a0bdc5c7572\") " Nov 23 07:02:06 crc kubenswrapper[4559]: I1123 07:02:06.253768 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lswtn\" (UniqueName: \"kubernetes.io/projected/27bbb5a9-7e24-4529-a05e-1a0bdc5c7572-kube-api-access-lswtn\") pod \"27bbb5a9-7e24-4529-a05e-1a0bdc5c7572\" (UID: \"27bbb5a9-7e24-4529-a05e-1a0bdc5c7572\") " Nov 23 07:02:06 crc kubenswrapper[4559]: I1123 07:02:06.259998 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27bbb5a9-7e24-4529-a05e-1a0bdc5c7572-kube-api-access-lswtn" (OuterVolumeSpecName: "kube-api-access-lswtn") pod "27bbb5a9-7e24-4529-a05e-1a0bdc5c7572" (UID: "27bbb5a9-7e24-4529-a05e-1a0bdc5c7572"). InnerVolumeSpecName "kube-api-access-lswtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:02:06 crc kubenswrapper[4559]: I1123 07:02:06.282635 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27bbb5a9-7e24-4529-a05e-1a0bdc5c7572-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "27bbb5a9-7e24-4529-a05e-1a0bdc5c7572" (UID: "27bbb5a9-7e24-4529-a05e-1a0bdc5c7572"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:02:06 crc kubenswrapper[4559]: I1123 07:02:06.282688 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27bbb5a9-7e24-4529-a05e-1a0bdc5c7572-config-data" (OuterVolumeSpecName: "config-data") pod "27bbb5a9-7e24-4529-a05e-1a0bdc5c7572" (UID: "27bbb5a9-7e24-4529-a05e-1a0bdc5c7572"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:02:06 crc kubenswrapper[4559]: I1123 07:02:06.359180 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27bbb5a9-7e24-4529-a05e-1a0bdc5c7572-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:06 crc kubenswrapper[4559]: I1123 07:02:06.359212 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lswtn\" (UniqueName: \"kubernetes.io/projected/27bbb5a9-7e24-4529-a05e-1a0bdc5c7572-kube-api-access-lswtn\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:06 crc kubenswrapper[4559]: I1123 07:02:06.359224 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27bbb5a9-7e24-4529-a05e-1a0bdc5c7572-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:07 crc kubenswrapper[4559]: I1123 07:02:07.042338 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"27bbb5a9-7e24-4529-a05e-1a0bdc5c7572","Type":"ContainerDied","Data":"9328cf35b5c593f87eebdf648275dde55e2b3b4cf8ba43cc8d1775bdd21c3f19"} Nov 23 07:02:07 crc kubenswrapper[4559]: I1123 07:02:07.042673 4559 scope.go:117] "RemoveContainer" containerID="6eff3da2d979d6fd24ecae38d241fea26cc86e4a77ec0a6676a5d6e92fdc15a4" Nov 23 07:02:07 crc kubenswrapper[4559]: I1123 07:02:07.042951 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 07:02:07 crc kubenswrapper[4559]: I1123 07:02:07.059631 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:02:07 crc kubenswrapper[4559]: I1123 07:02:07.066098 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:02:07 crc kubenswrapper[4559]: I1123 07:02:07.077223 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:02:07 crc kubenswrapper[4559]: E1123 07:02:07.077687 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27bbb5a9-7e24-4529-a05e-1a0bdc5c7572" containerName="nova-scheduler-scheduler" Nov 23 07:02:07 crc kubenswrapper[4559]: I1123 07:02:07.077753 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="27bbb5a9-7e24-4529-a05e-1a0bdc5c7572" containerName="nova-scheduler-scheduler" Nov 23 07:02:07 crc kubenswrapper[4559]: E1123 07:02:07.077806 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c818ed2a-d851-4a13-b67d-967f178654dc" containerName="nova-manage" Nov 23 07:02:07 crc kubenswrapper[4559]: I1123 07:02:07.077873 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="c818ed2a-d851-4a13-b67d-967f178654dc" containerName="nova-manage" Nov 23 07:02:07 crc kubenswrapper[4559]: I1123 07:02:07.078287 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="c818ed2a-d851-4a13-b67d-967f178654dc" containerName="nova-manage" Nov 23 07:02:07 crc kubenswrapper[4559]: I1123 07:02:07.078384 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="27bbb5a9-7e24-4529-a05e-1a0bdc5c7572" containerName="nova-scheduler-scheduler" Nov 23 07:02:07 crc kubenswrapper[4559]: I1123 07:02:07.079035 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 07:02:07 crc kubenswrapper[4559]: I1123 07:02:07.081961 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 23 07:02:07 crc kubenswrapper[4559]: I1123 07:02:07.092913 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:02:07 crc kubenswrapper[4559]: I1123 07:02:07.172384 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwlxc\" (UniqueName: \"kubernetes.io/projected/ba80f653-69d6-4a80-aaab-2747e0f9b61b-kube-api-access-pwlxc\") pod \"nova-scheduler-0\" (UID: \"ba80f653-69d6-4a80-aaab-2747e0f9b61b\") " pod="openstack/nova-scheduler-0" Nov 23 07:02:07 crc kubenswrapper[4559]: I1123 07:02:07.172442 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba80f653-69d6-4a80-aaab-2747e0f9b61b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ba80f653-69d6-4a80-aaab-2747e0f9b61b\") " pod="openstack/nova-scheduler-0" Nov 23 07:02:07 crc kubenswrapper[4559]: I1123 07:02:07.172551 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba80f653-69d6-4a80-aaab-2747e0f9b61b-config-data\") pod \"nova-scheduler-0\" (UID: \"ba80f653-69d6-4a80-aaab-2747e0f9b61b\") " pod="openstack/nova-scheduler-0" Nov 23 07:02:07 crc kubenswrapper[4559]: I1123 07:02:07.274924 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwlxc\" (UniqueName: \"kubernetes.io/projected/ba80f653-69d6-4a80-aaab-2747e0f9b61b-kube-api-access-pwlxc\") pod \"nova-scheduler-0\" (UID: \"ba80f653-69d6-4a80-aaab-2747e0f9b61b\") " pod="openstack/nova-scheduler-0" Nov 23 07:02:07 crc kubenswrapper[4559]: I1123 07:02:07.275003 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba80f653-69d6-4a80-aaab-2747e0f9b61b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ba80f653-69d6-4a80-aaab-2747e0f9b61b\") " pod="openstack/nova-scheduler-0" Nov 23 07:02:07 crc kubenswrapper[4559]: I1123 07:02:07.275155 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba80f653-69d6-4a80-aaab-2747e0f9b61b-config-data\") pod \"nova-scheduler-0\" (UID: \"ba80f653-69d6-4a80-aaab-2747e0f9b61b\") " pod="openstack/nova-scheduler-0" Nov 23 07:02:07 crc kubenswrapper[4559]: I1123 07:02:07.280114 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba80f653-69d6-4a80-aaab-2747e0f9b61b-config-data\") pod \"nova-scheduler-0\" (UID: \"ba80f653-69d6-4a80-aaab-2747e0f9b61b\") " pod="openstack/nova-scheduler-0" Nov 23 07:02:07 crc kubenswrapper[4559]: I1123 07:02:07.280688 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba80f653-69d6-4a80-aaab-2747e0f9b61b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ba80f653-69d6-4a80-aaab-2747e0f9b61b\") " pod="openstack/nova-scheduler-0" Nov 23 07:02:07 crc kubenswrapper[4559]: I1123 07:02:07.289015 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwlxc\" (UniqueName: \"kubernetes.io/projected/ba80f653-69d6-4a80-aaab-2747e0f9b61b-kube-api-access-pwlxc\") pod \"nova-scheduler-0\" (UID: \"ba80f653-69d6-4a80-aaab-2747e0f9b61b\") " pod="openstack/nova-scheduler-0" Nov 23 07:02:07 crc kubenswrapper[4559]: I1123 07:02:07.392540 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 07:02:07 crc kubenswrapper[4559]: I1123 07:02:07.780444 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 07:02:07 crc kubenswrapper[4559]: W1123 07:02:07.785074 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba80f653_69d6_4a80_aaab_2747e0f9b61b.slice/crio-a0e04a829246007d28f19778b2fc9c8d3595aff80f5e751d961b5e42349c45bc WatchSource:0}: Error finding container a0e04a829246007d28f19778b2fc9c8d3595aff80f5e751d961b5e42349c45bc: Status 404 returned error can't find the container with id a0e04a829246007d28f19778b2fc9c8d3595aff80f5e751d961b5e42349c45bc Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.057406 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ba80f653-69d6-4a80-aaab-2747e0f9b61b","Type":"ContainerStarted","Data":"f455293c6b0f2bb8d11b2b38110d81c2a7360a4afc179b950b8cd0d1cdf645bb"} Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.057454 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ba80f653-69d6-4a80-aaab-2747e0f9b61b","Type":"ContainerStarted","Data":"a0e04a829246007d28f19778b2fc9c8d3595aff80f5e751d961b5e42349c45bc"} Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.079746 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.079723175 podStartE2EDuration="1.079723175s" podCreationTimestamp="2025-11-23 07:02:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:02:08.069815653 +0000 UTC m=+1030.091801267" watchObservedRunningTime="2025-11-23 07:02:08.079723175 +0000 UTC m=+1030.101708790" Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.300154 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27bbb5a9-7e24-4529-a05e-1a0bdc5c7572" path="/var/lib/kubelet/pods/27bbb5a9-7e24-4529-a05e-1a0bdc5c7572/volumes" Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.703222 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.707442 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.810011 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a4a109b-31a8-42a0-a025-05652e26b669-internal-tls-certs\") pod \"8a4a109b-31a8-42a0-a025-05652e26b669\" (UID: \"8a4a109b-31a8-42a0-a025-05652e26b669\") " Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.810093 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7n5n\" (UniqueName: \"kubernetes.io/projected/8d0d4e86-3876-4666-b628-83cb6c08700b-kube-api-access-p7n5n\") pod \"8d0d4e86-3876-4666-b628-83cb6c08700b\" (UID: \"8d0d4e86-3876-4666-b628-83cb6c08700b\") " Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.810132 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d0d4e86-3876-4666-b628-83cb6c08700b-logs\") pod \"8d0d4e86-3876-4666-b628-83cb6c08700b\" (UID: \"8d0d4e86-3876-4666-b628-83cb6c08700b\") " Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.810163 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a4a109b-31a8-42a0-a025-05652e26b669-combined-ca-bundle\") pod \"8a4a109b-31a8-42a0-a025-05652e26b669\" (UID: \"8a4a109b-31a8-42a0-a025-05652e26b669\") " Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.810253 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d0d4e86-3876-4666-b628-83cb6c08700b-config-data\") pod \"8d0d4e86-3876-4666-b628-83cb6c08700b\" (UID: \"8d0d4e86-3876-4666-b628-83cb6c08700b\") " Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.810298 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a4a109b-31a8-42a0-a025-05652e26b669-config-data\") pod \"8a4a109b-31a8-42a0-a025-05652e26b669\" (UID: \"8a4a109b-31a8-42a0-a025-05652e26b669\") " Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.810373 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d0d4e86-3876-4666-b628-83cb6c08700b-combined-ca-bundle\") pod \"8d0d4e86-3876-4666-b628-83cb6c08700b\" (UID: \"8d0d4e86-3876-4666-b628-83cb6c08700b\") " Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.810524 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a4a109b-31a8-42a0-a025-05652e26b669-public-tls-certs\") pod \"8a4a109b-31a8-42a0-a025-05652e26b669\" (UID: \"8a4a109b-31a8-42a0-a025-05652e26b669\") " Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.810558 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cg9s\" (UniqueName: \"kubernetes.io/projected/8a4a109b-31a8-42a0-a025-05652e26b669-kube-api-access-7cg9s\") pod \"8a4a109b-31a8-42a0-a025-05652e26b669\" (UID: \"8a4a109b-31a8-42a0-a025-05652e26b669\") " Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.810590 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a4a109b-31a8-42a0-a025-05652e26b669-logs\") pod \"8a4a109b-31a8-42a0-a025-05652e26b669\" (UID: \"8a4a109b-31a8-42a0-a025-05652e26b669\") " Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.810611 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d0d4e86-3876-4666-b628-83cb6c08700b-nova-metadata-tls-certs\") pod \"8d0d4e86-3876-4666-b628-83cb6c08700b\" (UID: \"8d0d4e86-3876-4666-b628-83cb6c08700b\") " Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.810861 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d0d4e86-3876-4666-b628-83cb6c08700b-logs" (OuterVolumeSpecName: "logs") pod "8d0d4e86-3876-4666-b628-83cb6c08700b" (UID: "8d0d4e86-3876-4666-b628-83cb6c08700b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.811235 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a4a109b-31a8-42a0-a025-05652e26b669-logs" (OuterVolumeSpecName: "logs") pod "8a4a109b-31a8-42a0-a025-05652e26b669" (UID: "8a4a109b-31a8-42a0-a025-05652e26b669"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.811794 4559 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a4a109b-31a8-42a0-a025-05652e26b669-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.811818 4559 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d0d4e86-3876-4666-b628-83cb6c08700b-logs\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.816066 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d0d4e86-3876-4666-b628-83cb6c08700b-kube-api-access-p7n5n" (OuterVolumeSpecName: "kube-api-access-p7n5n") pod "8d0d4e86-3876-4666-b628-83cb6c08700b" (UID: "8d0d4e86-3876-4666-b628-83cb6c08700b"). InnerVolumeSpecName "kube-api-access-p7n5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.817838 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a4a109b-31a8-42a0-a025-05652e26b669-kube-api-access-7cg9s" (OuterVolumeSpecName: "kube-api-access-7cg9s") pod "8a4a109b-31a8-42a0-a025-05652e26b669" (UID: "8a4a109b-31a8-42a0-a025-05652e26b669"). InnerVolumeSpecName "kube-api-access-7cg9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.836069 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d0d4e86-3876-4666-b628-83cb6c08700b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d0d4e86-3876-4666-b628-83cb6c08700b" (UID: "8d0d4e86-3876-4666-b628-83cb6c08700b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.839476 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d0d4e86-3876-4666-b628-83cb6c08700b-config-data" (OuterVolumeSpecName: "config-data") pod "8d0d4e86-3876-4666-b628-83cb6c08700b" (UID: "8d0d4e86-3876-4666-b628-83cb6c08700b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.840290 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a4a109b-31a8-42a0-a025-05652e26b669-config-data" (OuterVolumeSpecName: "config-data") pod "8a4a109b-31a8-42a0-a025-05652e26b669" (UID: "8a4a109b-31a8-42a0-a025-05652e26b669"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.845586 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a4a109b-31a8-42a0-a025-05652e26b669-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a4a109b-31a8-42a0-a025-05652e26b669" (UID: "8a4a109b-31a8-42a0-a025-05652e26b669"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.863172 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a4a109b-31a8-42a0-a025-05652e26b669-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8a4a109b-31a8-42a0-a025-05652e26b669" (UID: "8a4a109b-31a8-42a0-a025-05652e26b669"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.866807 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d0d4e86-3876-4666-b628-83cb6c08700b-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "8d0d4e86-3876-4666-b628-83cb6c08700b" (UID: "8d0d4e86-3876-4666-b628-83cb6c08700b"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.873622 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a4a109b-31a8-42a0-a025-05652e26b669-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8a4a109b-31a8-42a0-a025-05652e26b669" (UID: "8a4a109b-31a8-42a0-a025-05652e26b669"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.914031 4559 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a4a109b-31a8-42a0-a025-05652e26b669-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.914154 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cg9s\" (UniqueName: \"kubernetes.io/projected/8a4a109b-31a8-42a0-a025-05652e26b669-kube-api-access-7cg9s\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.914216 4559 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d0d4e86-3876-4666-b628-83cb6c08700b-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.914267 4559 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a4a109b-31a8-42a0-a025-05652e26b669-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.914317 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7n5n\" (UniqueName: \"kubernetes.io/projected/8d0d4e86-3876-4666-b628-83cb6c08700b-kube-api-access-p7n5n\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.914379 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a4a109b-31a8-42a0-a025-05652e26b669-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.914440 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d0d4e86-3876-4666-b628-83cb6c08700b-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.914495 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a4a109b-31a8-42a0-a025-05652e26b669-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:08 crc kubenswrapper[4559]: I1123 07:02:08.914544 4559 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d0d4e86-3876-4666-b628-83cb6c08700b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.066775 4559 generic.go:334] "Generic (PLEG): container finished" podID="8d0d4e86-3876-4666-b628-83cb6c08700b" containerID="9b5836b9d27c68a11951faf7967525bf1cd78e4a672b484c343e84c3363bac02" exitCode=0 Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.066832 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.066841 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8d0d4e86-3876-4666-b628-83cb6c08700b","Type":"ContainerDied","Data":"9b5836b9d27c68a11951faf7967525bf1cd78e4a672b484c343e84c3363bac02"} Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.067663 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8d0d4e86-3876-4666-b628-83cb6c08700b","Type":"ContainerDied","Data":"9882b4ab9934b0c3bf41b995a669faf80fc2133ae12977ecaa8cb978f65d6754"} Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.067689 4559 scope.go:117] "RemoveContainer" containerID="9b5836b9d27c68a11951faf7967525bf1cd78e4a672b484c343e84c3363bac02" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.069691 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.069712 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8a4a109b-31a8-42a0-a025-05652e26b669","Type":"ContainerDied","Data":"2b4ac7acd62790e8aea151ad93fa0fe2bfc3459660cc3f7946b8911783aa556d"} Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.069689 4559 generic.go:334] "Generic (PLEG): container finished" podID="8a4a109b-31a8-42a0-a025-05652e26b669" containerID="2b4ac7acd62790e8aea151ad93fa0fe2bfc3459660cc3f7946b8911783aa556d" exitCode=0 Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.069865 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8a4a109b-31a8-42a0-a025-05652e26b669","Type":"ContainerDied","Data":"04713f5599ce16ae416fc50fe4f76e93c329e50fad646746e4a80037acbf6dde"} Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.086462 4559 scope.go:117] "RemoveContainer" containerID="022d9852ba23da5dec14c28273df4974709ac8978ff19f432268af796578b513" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.103772 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.114250 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.114955 4559 scope.go:117] "RemoveContainer" containerID="9b5836b9d27c68a11951faf7967525bf1cd78e4a672b484c343e84c3363bac02" Nov 23 07:02:09 crc kubenswrapper[4559]: E1123 07:02:09.115368 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b5836b9d27c68a11951faf7967525bf1cd78e4a672b484c343e84c3363bac02\": container with ID starting with 9b5836b9d27c68a11951faf7967525bf1cd78e4a672b484c343e84c3363bac02 not found: ID does not exist" containerID="9b5836b9d27c68a11951faf7967525bf1cd78e4a672b484c343e84c3363bac02" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.115400 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b5836b9d27c68a11951faf7967525bf1cd78e4a672b484c343e84c3363bac02"} err="failed to get container status \"9b5836b9d27c68a11951faf7967525bf1cd78e4a672b484c343e84c3363bac02\": rpc error: code = NotFound desc = could not find container \"9b5836b9d27c68a11951faf7967525bf1cd78e4a672b484c343e84c3363bac02\": container with ID starting with 9b5836b9d27c68a11951faf7967525bf1cd78e4a672b484c343e84c3363bac02 not found: ID does not exist" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.115423 4559 scope.go:117] "RemoveContainer" containerID="022d9852ba23da5dec14c28273df4974709ac8978ff19f432268af796578b513" Nov 23 07:02:09 crc kubenswrapper[4559]: E1123 07:02:09.115664 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"022d9852ba23da5dec14c28273df4974709ac8978ff19f432268af796578b513\": container with ID starting with 022d9852ba23da5dec14c28273df4974709ac8978ff19f432268af796578b513 not found: ID does not exist" containerID="022d9852ba23da5dec14c28273df4974709ac8978ff19f432268af796578b513" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.115685 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"022d9852ba23da5dec14c28273df4974709ac8978ff19f432268af796578b513"} err="failed to get container status \"022d9852ba23da5dec14c28273df4974709ac8978ff19f432268af796578b513\": rpc error: code = NotFound desc = could not find container \"022d9852ba23da5dec14c28273df4974709ac8978ff19f432268af796578b513\": container with ID starting with 022d9852ba23da5dec14c28273df4974709ac8978ff19f432268af796578b513 not found: ID does not exist" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.115700 4559 scope.go:117] "RemoveContainer" containerID="2b4ac7acd62790e8aea151ad93fa0fe2bfc3459660cc3f7946b8911783aa556d" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.132043 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.138790 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.140401 4559 scope.go:117] "RemoveContainer" containerID="b0fdf7c9abb4776f1a0d53e9dc3458a5f19f90608d1edf65d82f2b3aa8bb6185" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.149267 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:02:09 crc kubenswrapper[4559]: E1123 07:02:09.149713 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d0d4e86-3876-4666-b628-83cb6c08700b" containerName="nova-metadata-log" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.149732 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d0d4e86-3876-4666-b628-83cb6c08700b" containerName="nova-metadata-log" Nov 23 07:02:09 crc kubenswrapper[4559]: E1123 07:02:09.149754 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d0d4e86-3876-4666-b628-83cb6c08700b" containerName="nova-metadata-metadata" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.149760 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d0d4e86-3876-4666-b628-83cb6c08700b" containerName="nova-metadata-metadata" Nov 23 07:02:09 crc kubenswrapper[4559]: E1123 07:02:09.149770 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a4a109b-31a8-42a0-a025-05652e26b669" containerName="nova-api-api" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.149776 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a4a109b-31a8-42a0-a025-05652e26b669" containerName="nova-api-api" Nov 23 07:02:09 crc kubenswrapper[4559]: E1123 07:02:09.149797 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a4a109b-31a8-42a0-a025-05652e26b669" containerName="nova-api-log" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.149802 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a4a109b-31a8-42a0-a025-05652e26b669" containerName="nova-api-log" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.150006 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d0d4e86-3876-4666-b628-83cb6c08700b" containerName="nova-metadata-log" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.150026 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a4a109b-31a8-42a0-a025-05652e26b669" containerName="nova-api-log" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.150032 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d0d4e86-3876-4666-b628-83cb6c08700b" containerName="nova-metadata-metadata" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.150052 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a4a109b-31a8-42a0-a025-05652e26b669" containerName="nova-api-api" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.150890 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.150974 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.154202 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.154283 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.155485 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.156653 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.157582 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.158151 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.158310 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.160371 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.162198 4559 scope.go:117] "RemoveContainer" containerID="2b4ac7acd62790e8aea151ad93fa0fe2bfc3459660cc3f7946b8911783aa556d" Nov 23 07:02:09 crc kubenswrapper[4559]: E1123 07:02:09.162611 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b4ac7acd62790e8aea151ad93fa0fe2bfc3459660cc3f7946b8911783aa556d\": container with ID starting with 2b4ac7acd62790e8aea151ad93fa0fe2bfc3459660cc3f7946b8911783aa556d not found: ID does not exist" containerID="2b4ac7acd62790e8aea151ad93fa0fe2bfc3459660cc3f7946b8911783aa556d" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.162669 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b4ac7acd62790e8aea151ad93fa0fe2bfc3459660cc3f7946b8911783aa556d"} err="failed to get container status \"2b4ac7acd62790e8aea151ad93fa0fe2bfc3459660cc3f7946b8911783aa556d\": rpc error: code = NotFound desc = could not find container \"2b4ac7acd62790e8aea151ad93fa0fe2bfc3459660cc3f7946b8911783aa556d\": container with ID starting with 2b4ac7acd62790e8aea151ad93fa0fe2bfc3459660cc3f7946b8911783aa556d not found: ID does not exist" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.162690 4559 scope.go:117] "RemoveContainer" containerID="b0fdf7c9abb4776f1a0d53e9dc3458a5f19f90608d1edf65d82f2b3aa8bb6185" Nov 23 07:02:09 crc kubenswrapper[4559]: E1123 07:02:09.163081 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0fdf7c9abb4776f1a0d53e9dc3458a5f19f90608d1edf65d82f2b3aa8bb6185\": container with ID starting with b0fdf7c9abb4776f1a0d53e9dc3458a5f19f90608d1edf65d82f2b3aa8bb6185 not found: ID does not exist" containerID="b0fdf7c9abb4776f1a0d53e9dc3458a5f19f90608d1edf65d82f2b3aa8bb6185" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.163121 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0fdf7c9abb4776f1a0d53e9dc3458a5f19f90608d1edf65d82f2b3aa8bb6185"} err="failed to get container status \"b0fdf7c9abb4776f1a0d53e9dc3458a5f19f90608d1edf65d82f2b3aa8bb6185\": rpc error: code = NotFound desc = could not find container \"b0fdf7c9abb4776f1a0d53e9dc3458a5f19f90608d1edf65d82f2b3aa8bb6185\": container with ID starting with b0fdf7c9abb4776f1a0d53e9dc3458a5f19f90608d1edf65d82f2b3aa8bb6185 not found: ID does not exist" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.328105 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/66d6e144-f771-40b4-a6f2-0689654cf797-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"66d6e144-f771-40b4-a6f2-0689654cf797\") " pod="openstack/nova-metadata-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.328352 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e55ca2b9-735d-434f-899d-cef9bce42b2e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e55ca2b9-735d-434f-899d-cef9bce42b2e\") " pod="openstack/nova-api-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.328469 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e55ca2b9-735d-434f-899d-cef9bce42b2e-config-data\") pod \"nova-api-0\" (UID: \"e55ca2b9-735d-434f-899d-cef9bce42b2e\") " pod="openstack/nova-api-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.328608 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjpgh\" (UniqueName: \"kubernetes.io/projected/e55ca2b9-735d-434f-899d-cef9bce42b2e-kube-api-access-fjpgh\") pod \"nova-api-0\" (UID: \"e55ca2b9-735d-434f-899d-cef9bce42b2e\") " pod="openstack/nova-api-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.328762 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e55ca2b9-735d-434f-899d-cef9bce42b2e-public-tls-certs\") pod \"nova-api-0\" (UID: \"e55ca2b9-735d-434f-899d-cef9bce42b2e\") " pod="openstack/nova-api-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.328840 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e55ca2b9-735d-434f-899d-cef9bce42b2e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e55ca2b9-735d-434f-899d-cef9bce42b2e\") " pod="openstack/nova-api-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.329074 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66d6e144-f771-40b4-a6f2-0689654cf797-config-data\") pod \"nova-metadata-0\" (UID: \"66d6e144-f771-40b4-a6f2-0689654cf797\") " pod="openstack/nova-metadata-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.329095 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66d6e144-f771-40b4-a6f2-0689654cf797-logs\") pod \"nova-metadata-0\" (UID: \"66d6e144-f771-40b4-a6f2-0689654cf797\") " pod="openstack/nova-metadata-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.329124 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b27hq\" (UniqueName: \"kubernetes.io/projected/66d6e144-f771-40b4-a6f2-0689654cf797-kube-api-access-b27hq\") pod \"nova-metadata-0\" (UID: \"66d6e144-f771-40b4-a6f2-0689654cf797\") " pod="openstack/nova-metadata-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.329246 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e55ca2b9-735d-434f-899d-cef9bce42b2e-logs\") pod \"nova-api-0\" (UID: \"e55ca2b9-735d-434f-899d-cef9bce42b2e\") " pod="openstack/nova-api-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.329374 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66d6e144-f771-40b4-a6f2-0689654cf797-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"66d6e144-f771-40b4-a6f2-0689654cf797\") " pod="openstack/nova-metadata-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.431243 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/66d6e144-f771-40b4-a6f2-0689654cf797-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"66d6e144-f771-40b4-a6f2-0689654cf797\") " pod="openstack/nova-metadata-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.431287 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e55ca2b9-735d-434f-899d-cef9bce42b2e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e55ca2b9-735d-434f-899d-cef9bce42b2e\") " pod="openstack/nova-api-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.431309 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e55ca2b9-735d-434f-899d-cef9bce42b2e-config-data\") pod \"nova-api-0\" (UID: \"e55ca2b9-735d-434f-899d-cef9bce42b2e\") " pod="openstack/nova-api-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.431337 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjpgh\" (UniqueName: \"kubernetes.io/projected/e55ca2b9-735d-434f-899d-cef9bce42b2e-kube-api-access-fjpgh\") pod \"nova-api-0\" (UID: \"e55ca2b9-735d-434f-899d-cef9bce42b2e\") " pod="openstack/nova-api-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.431354 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e55ca2b9-735d-434f-899d-cef9bce42b2e-public-tls-certs\") pod \"nova-api-0\" (UID: \"e55ca2b9-735d-434f-899d-cef9bce42b2e\") " pod="openstack/nova-api-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.431380 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e55ca2b9-735d-434f-899d-cef9bce42b2e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e55ca2b9-735d-434f-899d-cef9bce42b2e\") " pod="openstack/nova-api-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.431473 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66d6e144-f771-40b4-a6f2-0689654cf797-config-data\") pod \"nova-metadata-0\" (UID: \"66d6e144-f771-40b4-a6f2-0689654cf797\") " pod="openstack/nova-metadata-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.431489 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66d6e144-f771-40b4-a6f2-0689654cf797-logs\") pod \"nova-metadata-0\" (UID: \"66d6e144-f771-40b4-a6f2-0689654cf797\") " pod="openstack/nova-metadata-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.431505 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b27hq\" (UniqueName: \"kubernetes.io/projected/66d6e144-f771-40b4-a6f2-0689654cf797-kube-api-access-b27hq\") pod \"nova-metadata-0\" (UID: \"66d6e144-f771-40b4-a6f2-0689654cf797\") " pod="openstack/nova-metadata-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.431557 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e55ca2b9-735d-434f-899d-cef9bce42b2e-logs\") pod \"nova-api-0\" (UID: \"e55ca2b9-735d-434f-899d-cef9bce42b2e\") " pod="openstack/nova-api-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.431608 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66d6e144-f771-40b4-a6f2-0689654cf797-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"66d6e144-f771-40b4-a6f2-0689654cf797\") " pod="openstack/nova-metadata-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.432278 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66d6e144-f771-40b4-a6f2-0689654cf797-logs\") pod \"nova-metadata-0\" (UID: \"66d6e144-f771-40b4-a6f2-0689654cf797\") " pod="openstack/nova-metadata-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.432720 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e55ca2b9-735d-434f-899d-cef9bce42b2e-logs\") pod \"nova-api-0\" (UID: \"e55ca2b9-735d-434f-899d-cef9bce42b2e\") " pod="openstack/nova-api-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.436138 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e55ca2b9-735d-434f-899d-cef9bce42b2e-config-data\") pod \"nova-api-0\" (UID: \"e55ca2b9-735d-434f-899d-cef9bce42b2e\") " pod="openstack/nova-api-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.436599 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/66d6e144-f771-40b4-a6f2-0689654cf797-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"66d6e144-f771-40b4-a6f2-0689654cf797\") " pod="openstack/nova-metadata-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.437051 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e55ca2b9-735d-434f-899d-cef9bce42b2e-public-tls-certs\") pod \"nova-api-0\" (UID: \"e55ca2b9-735d-434f-899d-cef9bce42b2e\") " pod="openstack/nova-api-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.437149 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e55ca2b9-735d-434f-899d-cef9bce42b2e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e55ca2b9-735d-434f-899d-cef9bce42b2e\") " pod="openstack/nova-api-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.438746 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e55ca2b9-735d-434f-899d-cef9bce42b2e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e55ca2b9-735d-434f-899d-cef9bce42b2e\") " pod="openstack/nova-api-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.440570 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66d6e144-f771-40b4-a6f2-0689654cf797-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"66d6e144-f771-40b4-a6f2-0689654cf797\") " pod="openstack/nova-metadata-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.445721 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b27hq\" (UniqueName: \"kubernetes.io/projected/66d6e144-f771-40b4-a6f2-0689654cf797-kube-api-access-b27hq\") pod \"nova-metadata-0\" (UID: \"66d6e144-f771-40b4-a6f2-0689654cf797\") " pod="openstack/nova-metadata-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.446258 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjpgh\" (UniqueName: \"kubernetes.io/projected/e55ca2b9-735d-434f-899d-cef9bce42b2e-kube-api-access-fjpgh\") pod \"nova-api-0\" (UID: \"e55ca2b9-735d-434f-899d-cef9bce42b2e\") " pod="openstack/nova-api-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.451973 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66d6e144-f771-40b4-a6f2-0689654cf797-config-data\") pod \"nova-metadata-0\" (UID: \"66d6e144-f771-40b4-a6f2-0689654cf797\") " pod="openstack/nova-metadata-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.474915 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.485988 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.856092 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 07:02:09 crc kubenswrapper[4559]: I1123 07:02:09.921918 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 23 07:02:09 crc kubenswrapper[4559]: W1123 07:02:09.930089 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode55ca2b9_735d_434f_899d_cef9bce42b2e.slice/crio-4ba0832c76f8d8757d91dd9897b4ebfdb2794bb1992cb755adf5657db79d1e69 WatchSource:0}: Error finding container 4ba0832c76f8d8757d91dd9897b4ebfdb2794bb1992cb755adf5657db79d1e69: Status 404 returned error can't find the container with id 4ba0832c76f8d8757d91dd9897b4ebfdb2794bb1992cb755adf5657db79d1e69 Nov 23 07:02:10 crc kubenswrapper[4559]: I1123 07:02:10.079057 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"66d6e144-f771-40b4-a6f2-0689654cf797","Type":"ContainerStarted","Data":"b3e6f738ed138bfebfdb1457068fdcd21e2a1faa7fbacf3691207f29e2143af7"} Nov 23 07:02:10 crc kubenswrapper[4559]: I1123 07:02:10.079103 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"66d6e144-f771-40b4-a6f2-0689654cf797","Type":"ContainerStarted","Data":"b8c00e8bf9b579de6ec7f216806de840a7d53cdc3cfe29b24c747aa66e6a6316"} Nov 23 07:02:10 crc kubenswrapper[4559]: I1123 07:02:10.080596 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e55ca2b9-735d-434f-899d-cef9bce42b2e","Type":"ContainerStarted","Data":"25cd1a2c50feb749cdda2ef313e1be0c51e3e16b4190f50d059e19f2e70d30e7"} Nov 23 07:02:10 crc kubenswrapper[4559]: I1123 07:02:10.080636 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e55ca2b9-735d-434f-899d-cef9bce42b2e","Type":"ContainerStarted","Data":"4ba0832c76f8d8757d91dd9897b4ebfdb2794bb1992cb755adf5657db79d1e69"} Nov 23 07:02:10 crc kubenswrapper[4559]: I1123 07:02:10.284021 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a4a109b-31a8-42a0-a025-05652e26b669" path="/var/lib/kubelet/pods/8a4a109b-31a8-42a0-a025-05652e26b669/volumes" Nov 23 07:02:10 crc kubenswrapper[4559]: I1123 07:02:10.285242 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d0d4e86-3876-4666-b628-83cb6c08700b" path="/var/lib/kubelet/pods/8d0d4e86-3876-4666-b628-83cb6c08700b/volumes" Nov 23 07:02:11 crc kubenswrapper[4559]: I1123 07:02:11.092011 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"66d6e144-f771-40b4-a6f2-0689654cf797","Type":"ContainerStarted","Data":"29cdf61fa1310bb6c0de241a865947221aff8fc2c9c15180d4957762a265e322"} Nov 23 07:02:11 crc kubenswrapper[4559]: I1123 07:02:11.094063 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e55ca2b9-735d-434f-899d-cef9bce42b2e","Type":"ContainerStarted","Data":"035956d19a2b17216f93303a9e7933390b3ee2112822b37d70121319b73ad564"} Nov 23 07:02:11 crc kubenswrapper[4559]: I1123 07:02:11.111758 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.111740061 podStartE2EDuration="2.111740061s" podCreationTimestamp="2025-11-23 07:02:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:02:11.105699782 +0000 UTC m=+1033.127685395" watchObservedRunningTime="2025-11-23 07:02:11.111740061 +0000 UTC m=+1033.133725674" Nov 23 07:02:11 crc kubenswrapper[4559]: I1123 07:02:11.125470 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.125452118 podStartE2EDuration="2.125452118s" podCreationTimestamp="2025-11-23 07:02:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:02:11.122160428 +0000 UTC m=+1033.144146052" watchObservedRunningTime="2025-11-23 07:02:11.125452118 +0000 UTC m=+1033.147437732" Nov 23 07:02:12 crc kubenswrapper[4559]: I1123 07:02:12.392719 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 23 07:02:14 crc kubenswrapper[4559]: I1123 07:02:14.475875 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 23 07:02:14 crc kubenswrapper[4559]: I1123 07:02:14.476113 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 23 07:02:17 crc kubenswrapper[4559]: I1123 07:02:17.392809 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 23 07:02:17 crc kubenswrapper[4559]: I1123 07:02:17.416965 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 23 07:02:18 crc kubenswrapper[4559]: I1123 07:02:18.170818 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 23 07:02:19 crc kubenswrapper[4559]: I1123 07:02:19.475364 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 23 07:02:19 crc kubenswrapper[4559]: I1123 07:02:19.475743 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 23 07:02:19 crc kubenswrapper[4559]: I1123 07:02:19.487476 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 23 07:02:19 crc kubenswrapper[4559]: I1123 07:02:19.487525 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 23 07:02:20 crc kubenswrapper[4559]: I1123 07:02:20.489772 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="66d6e144-f771-40b4-a6f2-0689654cf797" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 23 07:02:20 crc kubenswrapper[4559]: I1123 07:02:20.489788 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="66d6e144-f771-40b4-a6f2-0689654cf797" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 23 07:02:20 crc kubenswrapper[4559]: I1123 07:02:20.499767 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e55ca2b9-735d-434f-899d-cef9bce42b2e" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 23 07:02:20 crc kubenswrapper[4559]: I1123 07:02:20.499764 4559 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e55ca2b9-735d-434f-899d-cef9bce42b2e" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 23 07:02:21 crc kubenswrapper[4559]: I1123 07:02:21.190359 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 23 07:02:26 crc kubenswrapper[4559]: I1123 07:02:26.167066 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:02:26 crc kubenswrapper[4559]: I1123 07:02:26.167498 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:02:29 crc kubenswrapper[4559]: I1123 07:02:29.479998 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 23 07:02:29 crc kubenswrapper[4559]: I1123 07:02:29.480563 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 23 07:02:29 crc kubenswrapper[4559]: I1123 07:02:29.484834 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 23 07:02:29 crc kubenswrapper[4559]: I1123 07:02:29.485597 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 23 07:02:29 crc kubenswrapper[4559]: I1123 07:02:29.492495 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 23 07:02:29 crc kubenswrapper[4559]: I1123 07:02:29.492752 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 23 07:02:29 crc kubenswrapper[4559]: I1123 07:02:29.498157 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 23 07:02:29 crc kubenswrapper[4559]: I1123 07:02:29.502076 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 23 07:02:30 crc kubenswrapper[4559]: I1123 07:02:30.237240 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 23 07:02:30 crc kubenswrapper[4559]: I1123 07:02:30.242105 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 23 07:02:36 crc kubenswrapper[4559]: I1123 07:02:36.577272 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 23 07:02:37 crc kubenswrapper[4559]: I1123 07:02:37.222701 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 23 07:02:39 crc kubenswrapper[4559]: I1123 07:02:39.764893 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="a0fbc3f6-1c87-495a-a38f-5a0ae768ee20" containerName="rabbitmq" containerID="cri-o://059e75026cf507f704dab0c20ddda3da72f5483549bf88ab566bab645ad7eb61" gracePeriod=604797 Nov 23 07:02:40 crc kubenswrapper[4559]: I1123 07:02:40.448004 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="61e2f87a-db68-432a-8d13-94087d1afb17" containerName="rabbitmq" containerID="cri-o://94b662f9f4f42e326fff684a3b4eeacd903e6c78f20d5f630e934ee6cb875106" gracePeriod=604797 Nov 23 07:02:44 crc kubenswrapper[4559]: I1123 07:02:44.299134 4559 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="a0fbc3f6-1c87-495a-a38f-5a0ae768ee20" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.96:5671: connect: connection refused" Nov 23 07:02:44 crc kubenswrapper[4559]: I1123 07:02:44.564308 4559 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="61e2f87a-db68-432a-8d13-94087d1afb17" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.97:5671: connect: connection refused" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.139088 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.150848 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-rabbitmq-tls\") pod \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.150941 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-rabbitmq-plugins\") pod \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.150964 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zd2cg\" (UniqueName: \"kubernetes.io/projected/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-kube-api-access-zd2cg\") pod \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.150984 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.151087 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-rabbitmq-erlang-cookie\") pod \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.151130 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-rabbitmq-confd\") pod \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.151169 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-server-conf\") pod \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.151191 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-pod-info\") pod \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.151218 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-config-data\") pod \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.151234 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-plugins-conf\") pod \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.151250 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-erlang-cookie-secret\") pod \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\" (UID: \"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20\") " Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.151838 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "a0fbc3f6-1c87-495a-a38f-5a0ae768ee20" (UID: "a0fbc3f6-1c87-495a-a38f-5a0ae768ee20"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.151972 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "a0fbc3f6-1c87-495a-a38f-5a0ae768ee20" (UID: "a0fbc3f6-1c87-495a-a38f-5a0ae768ee20"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.152092 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "a0fbc3f6-1c87-495a-a38f-5a0ae768ee20" (UID: "a0fbc3f6-1c87-495a-a38f-5a0ae768ee20"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.152589 4559 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.152603 4559 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.152614 4559 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.157240 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-pod-info" (OuterVolumeSpecName: "pod-info") pod "a0fbc3f6-1c87-495a-a38f-5a0ae768ee20" (UID: "a0fbc3f6-1c87-495a-a38f-5a0ae768ee20"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.157780 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-kube-api-access-zd2cg" (OuterVolumeSpecName: "kube-api-access-zd2cg") pod "a0fbc3f6-1c87-495a-a38f-5a0ae768ee20" (UID: "a0fbc3f6-1c87-495a-a38f-5a0ae768ee20"). InnerVolumeSpecName "kube-api-access-zd2cg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.157856 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "a0fbc3f6-1c87-495a-a38f-5a0ae768ee20" (UID: "a0fbc3f6-1c87-495a-a38f-5a0ae768ee20"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.158161 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "a0fbc3f6-1c87-495a-a38f-5a0ae768ee20" (UID: "a0fbc3f6-1c87-495a-a38f-5a0ae768ee20"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.180742 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "a0fbc3f6-1c87-495a-a38f-5a0ae768ee20" (UID: "a0fbc3f6-1c87-495a-a38f-5a0ae768ee20"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.211338 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-config-data" (OuterVolumeSpecName: "config-data") pod "a0fbc3f6-1c87-495a-a38f-5a0ae768ee20" (UID: "a0fbc3f6-1c87-495a-a38f-5a0ae768ee20"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.224082 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-server-conf" (OuterVolumeSpecName: "server-conf") pod "a0fbc3f6-1c87-495a-a38f-5a0ae768ee20" (UID: "a0fbc3f6-1c87-495a-a38f-5a0ae768ee20"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.254116 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "a0fbc3f6-1c87-495a-a38f-5a0ae768ee20" (UID: "a0fbc3f6-1c87-495a-a38f-5a0ae768ee20"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.254323 4559 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.254352 4559 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-server-conf\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.254360 4559 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-pod-info\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.254370 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.254378 4559 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.254386 4559 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.254394 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zd2cg\" (UniqueName: \"kubernetes.io/projected/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20-kube-api-access-zd2cg\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.254429 4559 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.269375 4559 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.343378 4559 generic.go:334] "Generic (PLEG): container finished" podID="a0fbc3f6-1c87-495a-a38f-5a0ae768ee20" containerID="059e75026cf507f704dab0c20ddda3da72f5483549bf88ab566bab645ad7eb61" exitCode=0 Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.343426 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20","Type":"ContainerDied","Data":"059e75026cf507f704dab0c20ddda3da72f5483549bf88ab566bab645ad7eb61"} Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.343458 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a0fbc3f6-1c87-495a-a38f-5a0ae768ee20","Type":"ContainerDied","Data":"7458fff170fb3e48c0b29bbb98727c892d3f89f3718b88553d0f95c8c0752c96"} Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.343455 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.343477 4559 scope.go:117] "RemoveContainer" containerID="059e75026cf507f704dab0c20ddda3da72f5483549bf88ab566bab645ad7eb61" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.356154 4559 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.361693 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.367748 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.368663 4559 scope.go:117] "RemoveContainer" containerID="832a304e7da998cd94f31c322c8f84bedc01786f45ac24347d345e9f99aab1a6" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.389001 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 23 07:02:46 crc kubenswrapper[4559]: E1123 07:02:46.389379 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0fbc3f6-1c87-495a-a38f-5a0ae768ee20" containerName="setup-container" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.389395 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0fbc3f6-1c87-495a-a38f-5a0ae768ee20" containerName="setup-container" Nov 23 07:02:46 crc kubenswrapper[4559]: E1123 07:02:46.389414 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0fbc3f6-1c87-495a-a38f-5a0ae768ee20" containerName="rabbitmq" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.389420 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0fbc3f6-1c87-495a-a38f-5a0ae768ee20" containerName="rabbitmq" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.389597 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0fbc3f6-1c87-495a-a38f-5a0ae768ee20" containerName="rabbitmq" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.390931 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.393566 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.393723 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.393756 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.393805 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.393877 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-lgflk" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.393946 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.393967 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.398149 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.399395 4559 scope.go:117] "RemoveContainer" containerID="059e75026cf507f704dab0c20ddda3da72f5483549bf88ab566bab645ad7eb61" Nov 23 07:02:46 crc kubenswrapper[4559]: E1123 07:02:46.399796 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"059e75026cf507f704dab0c20ddda3da72f5483549bf88ab566bab645ad7eb61\": container with ID starting with 059e75026cf507f704dab0c20ddda3da72f5483549bf88ab566bab645ad7eb61 not found: ID does not exist" containerID="059e75026cf507f704dab0c20ddda3da72f5483549bf88ab566bab645ad7eb61" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.399831 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"059e75026cf507f704dab0c20ddda3da72f5483549bf88ab566bab645ad7eb61"} err="failed to get container status \"059e75026cf507f704dab0c20ddda3da72f5483549bf88ab566bab645ad7eb61\": rpc error: code = NotFound desc = could not find container \"059e75026cf507f704dab0c20ddda3da72f5483549bf88ab566bab645ad7eb61\": container with ID starting with 059e75026cf507f704dab0c20ddda3da72f5483549bf88ab566bab645ad7eb61 not found: ID does not exist" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.399856 4559 scope.go:117] "RemoveContainer" containerID="832a304e7da998cd94f31c322c8f84bedc01786f45ac24347d345e9f99aab1a6" Nov 23 07:02:46 crc kubenswrapper[4559]: E1123 07:02:46.400770 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"832a304e7da998cd94f31c322c8f84bedc01786f45ac24347d345e9f99aab1a6\": container with ID starting with 832a304e7da998cd94f31c322c8f84bedc01786f45ac24347d345e9f99aab1a6 not found: ID does not exist" containerID="832a304e7da998cd94f31c322c8f84bedc01786f45ac24347d345e9f99aab1a6" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.400808 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"832a304e7da998cd94f31c322c8f84bedc01786f45ac24347d345e9f99aab1a6"} err="failed to get container status \"832a304e7da998cd94f31c322c8f84bedc01786f45ac24347d345e9f99aab1a6\": rpc error: code = NotFound desc = could not find container \"832a304e7da998cd94f31c322c8f84bedc01786f45ac24347d345e9f99aab1a6\": container with ID starting with 832a304e7da998cd94f31c322c8f84bedc01786f45ac24347d345e9f99aab1a6 not found: ID does not exist" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.457436 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.457474 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.457507 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pd6m4\" (UniqueName: \"kubernetes.io/projected/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-kube-api-access-pd6m4\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.457617 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.457689 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.457717 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.457739 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.457819 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-config-data\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.458020 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.458073 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.458125 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.559466 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.559524 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.559573 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.559633 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.559667 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.559690 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pd6m4\" (UniqueName: \"kubernetes.io/projected/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-kube-api-access-pd6m4\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.559726 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.559743 4559 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.559747 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.560286 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.560308 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.560399 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-config-data\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.560492 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.560525 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.561052 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.561273 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.561296 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-config-data\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.563120 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.563162 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.563499 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.563994 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.575396 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pd6m4\" (UniqueName: \"kubernetes.io/projected/bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8-kube-api-access-pd6m4\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.581560 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8\") " pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.707372 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.738836 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.765673 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/61e2f87a-db68-432a-8d13-94087d1afb17-plugins-conf\") pod \"61e2f87a-db68-432a-8d13-94087d1afb17\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.765730 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/61e2f87a-db68-432a-8d13-94087d1afb17-rabbitmq-tls\") pod \"61e2f87a-db68-432a-8d13-94087d1afb17\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.765782 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/61e2f87a-db68-432a-8d13-94087d1afb17-server-conf\") pod \"61e2f87a-db68-432a-8d13-94087d1afb17\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.765833 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/61e2f87a-db68-432a-8d13-94087d1afb17-rabbitmq-erlang-cookie\") pod \"61e2f87a-db68-432a-8d13-94087d1afb17\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.765860 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"61e2f87a-db68-432a-8d13-94087d1afb17\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.765883 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9qg9\" (UniqueName: \"kubernetes.io/projected/61e2f87a-db68-432a-8d13-94087d1afb17-kube-api-access-s9qg9\") pod \"61e2f87a-db68-432a-8d13-94087d1afb17\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.765928 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/61e2f87a-db68-432a-8d13-94087d1afb17-pod-info\") pod \"61e2f87a-db68-432a-8d13-94087d1afb17\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.765956 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/61e2f87a-db68-432a-8d13-94087d1afb17-rabbitmq-confd\") pod \"61e2f87a-db68-432a-8d13-94087d1afb17\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.765991 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/61e2f87a-db68-432a-8d13-94087d1afb17-config-data\") pod \"61e2f87a-db68-432a-8d13-94087d1afb17\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.766048 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/61e2f87a-db68-432a-8d13-94087d1afb17-erlang-cookie-secret\") pod \"61e2f87a-db68-432a-8d13-94087d1afb17\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.766072 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/61e2f87a-db68-432a-8d13-94087d1afb17-rabbitmq-plugins\") pod \"61e2f87a-db68-432a-8d13-94087d1afb17\" (UID: \"61e2f87a-db68-432a-8d13-94087d1afb17\") " Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.766503 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61e2f87a-db68-432a-8d13-94087d1afb17-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "61e2f87a-db68-432a-8d13-94087d1afb17" (UID: "61e2f87a-db68-432a-8d13-94087d1afb17"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.766742 4559 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/61e2f87a-db68-432a-8d13-94087d1afb17-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.767159 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61e2f87a-db68-432a-8d13-94087d1afb17-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "61e2f87a-db68-432a-8d13-94087d1afb17" (UID: "61e2f87a-db68-432a-8d13-94087d1afb17"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.767561 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61e2f87a-db68-432a-8d13-94087d1afb17-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "61e2f87a-db68-432a-8d13-94087d1afb17" (UID: "61e2f87a-db68-432a-8d13-94087d1afb17"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.773549 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61e2f87a-db68-432a-8d13-94087d1afb17-kube-api-access-s9qg9" (OuterVolumeSpecName: "kube-api-access-s9qg9") pod "61e2f87a-db68-432a-8d13-94087d1afb17" (UID: "61e2f87a-db68-432a-8d13-94087d1afb17"). InnerVolumeSpecName "kube-api-access-s9qg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.773596 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61e2f87a-db68-432a-8d13-94087d1afb17-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "61e2f87a-db68-432a-8d13-94087d1afb17" (UID: "61e2f87a-db68-432a-8d13-94087d1afb17"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.774940 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/61e2f87a-db68-432a-8d13-94087d1afb17-pod-info" (OuterVolumeSpecName: "pod-info") pod "61e2f87a-db68-432a-8d13-94087d1afb17" (UID: "61e2f87a-db68-432a-8d13-94087d1afb17"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.790840 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61e2f87a-db68-432a-8d13-94087d1afb17-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "61e2f87a-db68-432a-8d13-94087d1afb17" (UID: "61e2f87a-db68-432a-8d13-94087d1afb17"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.790916 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "61e2f87a-db68-432a-8d13-94087d1afb17" (UID: "61e2f87a-db68-432a-8d13-94087d1afb17"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.791213 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61e2f87a-db68-432a-8d13-94087d1afb17-config-data" (OuterVolumeSpecName: "config-data") pod "61e2f87a-db68-432a-8d13-94087d1afb17" (UID: "61e2f87a-db68-432a-8d13-94087d1afb17"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.821028 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61e2f87a-db68-432a-8d13-94087d1afb17-server-conf" (OuterVolumeSpecName: "server-conf") pod "61e2f87a-db68-432a-8d13-94087d1afb17" (UID: "61e2f87a-db68-432a-8d13-94087d1afb17"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.868480 4559 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/61e2f87a-db68-432a-8d13-94087d1afb17-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.868511 4559 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/61e2f87a-db68-432a-8d13-94087d1afb17-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.868520 4559 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/61e2f87a-db68-432a-8d13-94087d1afb17-server-conf\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.868550 4559 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.868561 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9qg9\" (UniqueName: \"kubernetes.io/projected/61e2f87a-db68-432a-8d13-94087d1afb17-kube-api-access-s9qg9\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.868571 4559 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/61e2f87a-db68-432a-8d13-94087d1afb17-pod-info\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.868580 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/61e2f87a-db68-432a-8d13-94087d1afb17-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.868589 4559 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/61e2f87a-db68-432a-8d13-94087d1afb17-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.868596 4559 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/61e2f87a-db68-432a-8d13-94087d1afb17-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.882153 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61e2f87a-db68-432a-8d13-94087d1afb17-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "61e2f87a-db68-432a-8d13-94087d1afb17" (UID: "61e2f87a-db68-432a-8d13-94087d1afb17"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.891631 4559 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.969650 4559 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/61e2f87a-db68-432a-8d13-94087d1afb17-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:46 crc kubenswrapper[4559]: I1123 07:02:46.969680 4559 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.123034 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.350821 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8","Type":"ContainerStarted","Data":"86cc1801345e17b05b4fd311f0e71880d6e7c7385aeab1ce2101147a052cc693"} Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.352824 4559 generic.go:334] "Generic (PLEG): container finished" podID="61e2f87a-db68-432a-8d13-94087d1afb17" containerID="94b662f9f4f42e326fff684a3b4eeacd903e6c78f20d5f630e934ee6cb875106" exitCode=0 Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.352888 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"61e2f87a-db68-432a-8d13-94087d1afb17","Type":"ContainerDied","Data":"94b662f9f4f42e326fff684a3b4eeacd903e6c78f20d5f630e934ee6cb875106"} Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.352925 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"61e2f87a-db68-432a-8d13-94087d1afb17","Type":"ContainerDied","Data":"370d0ad9a0809d04c728aaedbc819ffca79441a16c588bc68e81736ed67ad218"} Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.352943 4559 scope.go:117] "RemoveContainer" containerID="94b662f9f4f42e326fff684a3b4eeacd903e6c78f20d5f630e934ee6cb875106" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.352942 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.369436 4559 scope.go:117] "RemoveContainer" containerID="3ed96cde7ed68fbff6a9ee5c583b3e101eecb01f52f5e3d5f2ca7b0c38bf7687" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.379674 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.390594 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.400244 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 23 07:02:47 crc kubenswrapper[4559]: E1123 07:02:47.400629 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61e2f87a-db68-432a-8d13-94087d1afb17" containerName="rabbitmq" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.400809 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="61e2f87a-db68-432a-8d13-94087d1afb17" containerName="rabbitmq" Nov 23 07:02:47 crc kubenswrapper[4559]: E1123 07:02:47.400853 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61e2f87a-db68-432a-8d13-94087d1afb17" containerName="setup-container" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.400861 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="61e2f87a-db68-432a-8d13-94087d1afb17" containerName="setup-container" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.401075 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="61e2f87a-db68-432a-8d13-94087d1afb17" containerName="rabbitmq" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.402037 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.403934 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.403945 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.404310 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.404421 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.404496 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.404555 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-frpw9" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.404436 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.417071 4559 scope.go:117] "RemoveContainer" containerID="94b662f9f4f42e326fff684a3b4eeacd903e6c78f20d5f630e934ee6cb875106" Nov 23 07:02:47 crc kubenswrapper[4559]: E1123 07:02:47.423073 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94b662f9f4f42e326fff684a3b4eeacd903e6c78f20d5f630e934ee6cb875106\": container with ID starting with 94b662f9f4f42e326fff684a3b4eeacd903e6c78f20d5f630e934ee6cb875106 not found: ID does not exist" containerID="94b662f9f4f42e326fff684a3b4eeacd903e6c78f20d5f630e934ee6cb875106" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.423110 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94b662f9f4f42e326fff684a3b4eeacd903e6c78f20d5f630e934ee6cb875106"} err="failed to get container status \"94b662f9f4f42e326fff684a3b4eeacd903e6c78f20d5f630e934ee6cb875106\": rpc error: code = NotFound desc = could not find container \"94b662f9f4f42e326fff684a3b4eeacd903e6c78f20d5f630e934ee6cb875106\": container with ID starting with 94b662f9f4f42e326fff684a3b4eeacd903e6c78f20d5f630e934ee6cb875106 not found: ID does not exist" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.423131 4559 scope.go:117] "RemoveContainer" containerID="3ed96cde7ed68fbff6a9ee5c583b3e101eecb01f52f5e3d5f2ca7b0c38bf7687" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.423333 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 23 07:02:47 crc kubenswrapper[4559]: E1123 07:02:47.423553 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ed96cde7ed68fbff6a9ee5c583b3e101eecb01f52f5e3d5f2ca7b0c38bf7687\": container with ID starting with 3ed96cde7ed68fbff6a9ee5c583b3e101eecb01f52f5e3d5f2ca7b0c38bf7687 not found: ID does not exist" containerID="3ed96cde7ed68fbff6a9ee5c583b3e101eecb01f52f5e3d5f2ca7b0c38bf7687" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.423576 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ed96cde7ed68fbff6a9ee5c583b3e101eecb01f52f5e3d5f2ca7b0c38bf7687"} err="failed to get container status \"3ed96cde7ed68fbff6a9ee5c583b3e101eecb01f52f5e3d5f2ca7b0c38bf7687\": rpc error: code = NotFound desc = could not find container \"3ed96cde7ed68fbff6a9ee5c583b3e101eecb01f52f5e3d5f2ca7b0c38bf7687\": container with ID starting with 3ed96cde7ed68fbff6a9ee5c583b3e101eecb01f52f5e3d5f2ca7b0c38bf7687 not found: ID does not exist" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.589177 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e0731be9-7300-45df-b028-8aeb2a3999b4-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.589222 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tt525\" (UniqueName: \"kubernetes.io/projected/e0731be9-7300-45df-b028-8aeb2a3999b4-kube-api-access-tt525\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.589260 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e0731be9-7300-45df-b028-8aeb2a3999b4-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.589278 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e0731be9-7300-45df-b028-8aeb2a3999b4-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.589386 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e0731be9-7300-45df-b028-8aeb2a3999b4-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.589456 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e0731be9-7300-45df-b028-8aeb2a3999b4-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.589634 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.589808 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e0731be9-7300-45df-b028-8aeb2a3999b4-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.589869 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e0731be9-7300-45df-b028-8aeb2a3999b4-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.589947 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e0731be9-7300-45df-b028-8aeb2a3999b4-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.589971 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e0731be9-7300-45df-b028-8aeb2a3999b4-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.692227 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.692333 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e0731be9-7300-45df-b028-8aeb2a3999b4-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.692373 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e0731be9-7300-45df-b028-8aeb2a3999b4-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.692423 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e0731be9-7300-45df-b028-8aeb2a3999b4-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.692443 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e0731be9-7300-45df-b028-8aeb2a3999b4-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.692520 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e0731be9-7300-45df-b028-8aeb2a3999b4-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.692545 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tt525\" (UniqueName: \"kubernetes.io/projected/e0731be9-7300-45df-b028-8aeb2a3999b4-kube-api-access-tt525\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.692564 4559 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.692572 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e0731be9-7300-45df-b028-8aeb2a3999b4-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.693172 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e0731be9-7300-45df-b028-8aeb2a3999b4-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.693308 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e0731be9-7300-45df-b028-8aeb2a3999b4-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.693564 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e0731be9-7300-45df-b028-8aeb2a3999b4-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.693592 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e0731be9-7300-45df-b028-8aeb2a3999b4-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.693839 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e0731be9-7300-45df-b028-8aeb2a3999b4-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.694354 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e0731be9-7300-45df-b028-8aeb2a3999b4-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.694917 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e0731be9-7300-45df-b028-8aeb2a3999b4-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.695458 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e0731be9-7300-45df-b028-8aeb2a3999b4-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.697782 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e0731be9-7300-45df-b028-8aeb2a3999b4-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.698203 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e0731be9-7300-45df-b028-8aeb2a3999b4-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.699516 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e0731be9-7300-45df-b028-8aeb2a3999b4-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.701131 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e0731be9-7300-45df-b028-8aeb2a3999b4-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.708298 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tt525\" (UniqueName: \"kubernetes.io/projected/e0731be9-7300-45df-b028-8aeb2a3999b4-kube-api-access-tt525\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.721225 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0731be9-7300-45df-b028-8aeb2a3999b4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.735429 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.928525 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cc6c548b7-pnb22"] Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.930828 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cc6c548b7-pnb22" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.937395 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.938655 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cc6c548b7-pnb22"] Nov 23 07:02:47 crc kubenswrapper[4559]: I1123 07:02:47.996593 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cc6c548b7-pnb22"] Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.007100 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-openstack-edpm-ipam\") pod \"dnsmasq-dns-7cc6c548b7-pnb22\" (UID: \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\") " pod="openstack/dnsmasq-dns-7cc6c548b7-pnb22" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.007137 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-dns-swift-storage-0\") pod \"dnsmasq-dns-7cc6c548b7-pnb22\" (UID: \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\") " pod="openstack/dnsmasq-dns-7cc6c548b7-pnb22" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.007159 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-ovsdbserver-nb\") pod \"dnsmasq-dns-7cc6c548b7-pnb22\" (UID: \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\") " pod="openstack/dnsmasq-dns-7cc6c548b7-pnb22" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.007217 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df7pq\" (UniqueName: \"kubernetes.io/projected/d5f99c59-ef7c-47fc-8902-4f291e65dc66-kube-api-access-df7pq\") pod \"dnsmasq-dns-7cc6c548b7-pnb22\" (UID: \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\") " pod="openstack/dnsmasq-dns-7cc6c548b7-pnb22" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.007273 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-ovsdbserver-sb\") pod \"dnsmasq-dns-7cc6c548b7-pnb22\" (UID: \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\") " pod="openstack/dnsmasq-dns-7cc6c548b7-pnb22" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.007406 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-dns-svc\") pod \"dnsmasq-dns-7cc6c548b7-pnb22\" (UID: \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\") " pod="openstack/dnsmasq-dns-7cc6c548b7-pnb22" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.007446 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-config\") pod \"dnsmasq-dns-7cc6c548b7-pnb22\" (UID: \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\") " pod="openstack/dnsmasq-dns-7cc6c548b7-pnb22" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.029768 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-869db7cd47-ghbfq"] Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.031291 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.042300 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-869db7cd47-ghbfq"] Nov 23 07:02:48 crc kubenswrapper[4559]: E1123 07:02:48.061135 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-df7pq openstack-edpm-ipam ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-7cc6c548b7-pnb22" podUID="d5f99c59-ef7c-47fc-8902-4f291e65dc66" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.106628 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.108867 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b236045-a74d-4a5e-a8e3-c70530d9e6a2-ovsdbserver-sb\") pod \"dnsmasq-dns-869db7cd47-ghbfq\" (UID: \"9b236045-a74d-4a5e-a8e3-c70530d9e6a2\") " pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.109233 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-config\") pod \"dnsmasq-dns-7cc6c548b7-pnb22\" (UID: \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\") " pod="openstack/dnsmasq-dns-7cc6c548b7-pnb22" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.109333 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b236045-a74d-4a5e-a8e3-c70530d9e6a2-dns-svc\") pod \"dnsmasq-dns-869db7cd47-ghbfq\" (UID: \"9b236045-a74d-4a5e-a8e3-c70530d9e6a2\") " pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.109368 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-dns-swift-storage-0\") pod \"dnsmasq-dns-7cc6c548b7-pnb22\" (UID: \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\") " pod="openstack/dnsmasq-dns-7cc6c548b7-pnb22" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.109384 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-openstack-edpm-ipam\") pod \"dnsmasq-dns-7cc6c548b7-pnb22\" (UID: \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\") " pod="openstack/dnsmasq-dns-7cc6c548b7-pnb22" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.109407 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-ovsdbserver-nb\") pod \"dnsmasq-dns-7cc6c548b7-pnb22\" (UID: \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\") " pod="openstack/dnsmasq-dns-7cc6c548b7-pnb22" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.109428 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9b236045-a74d-4a5e-a8e3-c70530d9e6a2-openstack-edpm-ipam\") pod \"dnsmasq-dns-869db7cd47-ghbfq\" (UID: \"9b236045-a74d-4a5e-a8e3-c70530d9e6a2\") " pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.109457 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df7pq\" (UniqueName: \"kubernetes.io/projected/d5f99c59-ef7c-47fc-8902-4f291e65dc66-kube-api-access-df7pq\") pod \"dnsmasq-dns-7cc6c548b7-pnb22\" (UID: \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\") " pod="openstack/dnsmasq-dns-7cc6c548b7-pnb22" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.109477 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b236045-a74d-4a5e-a8e3-c70530d9e6a2-config\") pod \"dnsmasq-dns-869db7cd47-ghbfq\" (UID: \"9b236045-a74d-4a5e-a8e3-c70530d9e6a2\") " pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.109513 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-ovsdbserver-sb\") pod \"dnsmasq-dns-7cc6c548b7-pnb22\" (UID: \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\") " pod="openstack/dnsmasq-dns-7cc6c548b7-pnb22" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.109535 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b236045-a74d-4a5e-a8e3-c70530d9e6a2-ovsdbserver-nb\") pod \"dnsmasq-dns-869db7cd47-ghbfq\" (UID: \"9b236045-a74d-4a5e-a8e3-c70530d9e6a2\") " pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.109561 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8xsn\" (UniqueName: \"kubernetes.io/projected/9b236045-a74d-4a5e-a8e3-c70530d9e6a2-kube-api-access-l8xsn\") pod \"dnsmasq-dns-869db7cd47-ghbfq\" (UID: \"9b236045-a74d-4a5e-a8e3-c70530d9e6a2\") " pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.109619 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9b236045-a74d-4a5e-a8e3-c70530d9e6a2-dns-swift-storage-0\") pod \"dnsmasq-dns-869db7cd47-ghbfq\" (UID: \"9b236045-a74d-4a5e-a8e3-c70530d9e6a2\") " pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.109695 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-dns-svc\") pod \"dnsmasq-dns-7cc6c548b7-pnb22\" (UID: \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\") " pod="openstack/dnsmasq-dns-7cc6c548b7-pnb22" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.110546 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-dns-svc\") pod \"dnsmasq-dns-7cc6c548b7-pnb22\" (UID: \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\") " pod="openstack/dnsmasq-dns-7cc6c548b7-pnb22" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.111353 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-ovsdbserver-nb\") pod \"dnsmasq-dns-7cc6c548b7-pnb22\" (UID: \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\") " pod="openstack/dnsmasq-dns-7cc6c548b7-pnb22" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.111444 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-config\") pod \"dnsmasq-dns-7cc6c548b7-pnb22\" (UID: \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\") " pod="openstack/dnsmasq-dns-7cc6c548b7-pnb22" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.111457 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-ovsdbserver-sb\") pod \"dnsmasq-dns-7cc6c548b7-pnb22\" (UID: \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\") " pod="openstack/dnsmasq-dns-7cc6c548b7-pnb22" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.112161 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-openstack-edpm-ipam\") pod \"dnsmasq-dns-7cc6c548b7-pnb22\" (UID: \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\") " pod="openstack/dnsmasq-dns-7cc6c548b7-pnb22" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.116498 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-dns-swift-storage-0\") pod \"dnsmasq-dns-7cc6c548b7-pnb22\" (UID: \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\") " pod="openstack/dnsmasq-dns-7cc6c548b7-pnb22" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.124261 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df7pq\" (UniqueName: \"kubernetes.io/projected/d5f99c59-ef7c-47fc-8902-4f291e65dc66-kube-api-access-df7pq\") pod \"dnsmasq-dns-7cc6c548b7-pnb22\" (UID: \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\") " pod="openstack/dnsmasq-dns-7cc6c548b7-pnb22" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.210840 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9b236045-a74d-4a5e-a8e3-c70530d9e6a2-dns-swift-storage-0\") pod \"dnsmasq-dns-869db7cd47-ghbfq\" (UID: \"9b236045-a74d-4a5e-a8e3-c70530d9e6a2\") " pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.210898 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b236045-a74d-4a5e-a8e3-c70530d9e6a2-ovsdbserver-sb\") pod \"dnsmasq-dns-869db7cd47-ghbfq\" (UID: \"9b236045-a74d-4a5e-a8e3-c70530d9e6a2\") " pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.210960 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b236045-a74d-4a5e-a8e3-c70530d9e6a2-dns-svc\") pod \"dnsmasq-dns-869db7cd47-ghbfq\" (UID: \"9b236045-a74d-4a5e-a8e3-c70530d9e6a2\") " pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.211001 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9b236045-a74d-4a5e-a8e3-c70530d9e6a2-openstack-edpm-ipam\") pod \"dnsmasq-dns-869db7cd47-ghbfq\" (UID: \"9b236045-a74d-4a5e-a8e3-c70530d9e6a2\") " pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.211048 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b236045-a74d-4a5e-a8e3-c70530d9e6a2-config\") pod \"dnsmasq-dns-869db7cd47-ghbfq\" (UID: \"9b236045-a74d-4a5e-a8e3-c70530d9e6a2\") " pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.211099 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b236045-a74d-4a5e-a8e3-c70530d9e6a2-ovsdbserver-nb\") pod \"dnsmasq-dns-869db7cd47-ghbfq\" (UID: \"9b236045-a74d-4a5e-a8e3-c70530d9e6a2\") " pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.211132 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8xsn\" (UniqueName: \"kubernetes.io/projected/9b236045-a74d-4a5e-a8e3-c70530d9e6a2-kube-api-access-l8xsn\") pod \"dnsmasq-dns-869db7cd47-ghbfq\" (UID: \"9b236045-a74d-4a5e-a8e3-c70530d9e6a2\") " pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.212047 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b236045-a74d-4a5e-a8e3-c70530d9e6a2-config\") pod \"dnsmasq-dns-869db7cd47-ghbfq\" (UID: \"9b236045-a74d-4a5e-a8e3-c70530d9e6a2\") " pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.212084 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b236045-a74d-4a5e-a8e3-c70530d9e6a2-dns-svc\") pod \"dnsmasq-dns-869db7cd47-ghbfq\" (UID: \"9b236045-a74d-4a5e-a8e3-c70530d9e6a2\") " pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.212187 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9b236045-a74d-4a5e-a8e3-c70530d9e6a2-openstack-edpm-ipam\") pod \"dnsmasq-dns-869db7cd47-ghbfq\" (UID: \"9b236045-a74d-4a5e-a8e3-c70530d9e6a2\") " pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.212089 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9b236045-a74d-4a5e-a8e3-c70530d9e6a2-dns-swift-storage-0\") pod \"dnsmasq-dns-869db7cd47-ghbfq\" (UID: \"9b236045-a74d-4a5e-a8e3-c70530d9e6a2\") " pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.212678 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b236045-a74d-4a5e-a8e3-c70530d9e6a2-ovsdbserver-nb\") pod \"dnsmasq-dns-869db7cd47-ghbfq\" (UID: \"9b236045-a74d-4a5e-a8e3-c70530d9e6a2\") " pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.214169 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b236045-a74d-4a5e-a8e3-c70530d9e6a2-ovsdbserver-sb\") pod \"dnsmasq-dns-869db7cd47-ghbfq\" (UID: \"9b236045-a74d-4a5e-a8e3-c70530d9e6a2\") " pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.224625 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8xsn\" (UniqueName: \"kubernetes.io/projected/9b236045-a74d-4a5e-a8e3-c70530d9e6a2-kube-api-access-l8xsn\") pod \"dnsmasq-dns-869db7cd47-ghbfq\" (UID: \"9b236045-a74d-4a5e-a8e3-c70530d9e6a2\") " pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.282704 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61e2f87a-db68-432a-8d13-94087d1afb17" path="/var/lib/kubelet/pods/61e2f87a-db68-432a-8d13-94087d1afb17/volumes" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.283421 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0fbc3f6-1c87-495a-a38f-5a0ae768ee20" path="/var/lib/kubelet/pods/a0fbc3f6-1c87-495a-a38f-5a0ae768ee20/volumes" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.362663 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e0731be9-7300-45df-b028-8aeb2a3999b4","Type":"ContainerStarted","Data":"b4908500f000469e8bc117e34cb50930f6aaed5786f70c4833a9169a5e9e79cb"} Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.364303 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8","Type":"ContainerStarted","Data":"2b13ac54b4bcbbd498556afe316007d3125b4fcde79e9bfef0769f0d7dff618e"} Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.366085 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cc6c548b7-pnb22" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.366999 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.374784 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cc6c548b7-pnb22" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.515422 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-config\") pod \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\" (UID: \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\") " Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.515741 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-dns-svc\") pod \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\" (UID: \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\") " Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.515812 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-ovsdbserver-sb\") pod \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\" (UID: \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\") " Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.515836 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-df7pq\" (UniqueName: \"kubernetes.io/projected/d5f99c59-ef7c-47fc-8902-4f291e65dc66-kube-api-access-df7pq\") pod \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\" (UID: \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\") " Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.515831 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-config" (OuterVolumeSpecName: "config") pod "d5f99c59-ef7c-47fc-8902-4f291e65dc66" (UID: "d5f99c59-ef7c-47fc-8902-4f291e65dc66"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.515851 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-dns-swift-storage-0\") pod \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\" (UID: \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\") " Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.515897 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-ovsdbserver-nb\") pod \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\" (UID: \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\") " Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.515912 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-openstack-edpm-ipam\") pod \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\" (UID: \"d5f99c59-ef7c-47fc-8902-4f291e65dc66\") " Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.516268 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d5f99c59-ef7c-47fc-8902-4f291e65dc66" (UID: "d5f99c59-ef7c-47fc-8902-4f291e65dc66"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.516466 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d5f99c59-ef7c-47fc-8902-4f291e65dc66" (UID: "d5f99c59-ef7c-47fc-8902-4f291e65dc66"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.516459 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d5f99c59-ef7c-47fc-8902-4f291e65dc66" (UID: "d5f99c59-ef7c-47fc-8902-4f291e65dc66"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.516575 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.516600 4559 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.516609 4559 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.516617 4559 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.516616 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "d5f99c59-ef7c-47fc-8902-4f291e65dc66" (UID: "d5f99c59-ef7c-47fc-8902-4f291e65dc66"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.516886 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d5f99c59-ef7c-47fc-8902-4f291e65dc66" (UID: "d5f99c59-ef7c-47fc-8902-4f291e65dc66"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.519931 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5f99c59-ef7c-47fc-8902-4f291e65dc66-kube-api-access-df7pq" (OuterVolumeSpecName: "kube-api-access-df7pq") pod "d5f99c59-ef7c-47fc-8902-4f291e65dc66" (UID: "d5f99c59-ef7c-47fc-8902-4f291e65dc66"). InnerVolumeSpecName "kube-api-access-df7pq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.618223 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-df7pq\" (UniqueName: \"kubernetes.io/projected/d5f99c59-ef7c-47fc-8902-4f291e65dc66-kube-api-access-df7pq\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.618248 4559 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:48 crc kubenswrapper[4559]: I1123 07:02:48.618258 4559 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5f99c59-ef7c-47fc-8902-4f291e65dc66-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:49 crc kubenswrapper[4559]: I1123 07:02:49.258140 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-869db7cd47-ghbfq"] Nov 23 07:02:49 crc kubenswrapper[4559]: W1123 07:02:49.261592 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b236045_a74d_4a5e_a8e3_c70530d9e6a2.slice/crio-3fec4ac7707c3c8aa78d055fcde56980ff9f1199b4e79987a25de51396e0a9ee WatchSource:0}: Error finding container 3fec4ac7707c3c8aa78d055fcde56980ff9f1199b4e79987a25de51396e0a9ee: Status 404 returned error can't find the container with id 3fec4ac7707c3c8aa78d055fcde56980ff9f1199b4e79987a25de51396e0a9ee Nov 23 07:02:49 crc kubenswrapper[4559]: I1123 07:02:49.374089 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e0731be9-7300-45df-b028-8aeb2a3999b4","Type":"ContainerStarted","Data":"60cd43dfdb5398753bd255ddfb13142a17a3642c8d822ebc86fdda896d0d281a"} Nov 23 07:02:49 crc kubenswrapper[4559]: I1123 07:02:49.375357 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cc6c548b7-pnb22" Nov 23 07:02:49 crc kubenswrapper[4559]: I1123 07:02:49.375433 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" event={"ID":"9b236045-a74d-4a5e-a8e3-c70530d9e6a2","Type":"ContainerStarted","Data":"3fec4ac7707c3c8aa78d055fcde56980ff9f1199b4e79987a25de51396e0a9ee"} Nov 23 07:02:49 crc kubenswrapper[4559]: I1123 07:02:49.432319 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cc6c548b7-pnb22"] Nov 23 07:02:49 crc kubenswrapper[4559]: I1123 07:02:49.438802 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cc6c548b7-pnb22"] Nov 23 07:02:50 crc kubenswrapper[4559]: I1123 07:02:50.281484 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5f99c59-ef7c-47fc-8902-4f291e65dc66" path="/var/lib/kubelet/pods/d5f99c59-ef7c-47fc-8902-4f291e65dc66/volumes" Nov 23 07:02:50 crc kubenswrapper[4559]: I1123 07:02:50.383853 4559 generic.go:334] "Generic (PLEG): container finished" podID="9b236045-a74d-4a5e-a8e3-c70530d9e6a2" containerID="7181c45e9293c4eca47afe7deeceb6bc4b4c0f18cac2ab77c1a92afa0b6e0fb3" exitCode=0 Nov 23 07:02:50 crc kubenswrapper[4559]: I1123 07:02:50.383924 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" event={"ID":"9b236045-a74d-4a5e-a8e3-c70530d9e6a2","Type":"ContainerDied","Data":"7181c45e9293c4eca47afe7deeceb6bc4b4c0f18cac2ab77c1a92afa0b6e0fb3"} Nov 23 07:02:51 crc kubenswrapper[4559]: I1123 07:02:51.393454 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" event={"ID":"9b236045-a74d-4a5e-a8e3-c70530d9e6a2","Type":"ContainerStarted","Data":"4f81acd02c983f6511f77f7cde618de52cf6d1fc03673ebf2d2c74c9b574e48a"} Nov 23 07:02:51 crc kubenswrapper[4559]: I1123 07:02:51.394293 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" Nov 23 07:02:51 crc kubenswrapper[4559]: I1123 07:02:51.408334 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" podStartSLOduration=3.40831584 podStartE2EDuration="3.40831584s" podCreationTimestamp="2025-11-23 07:02:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:02:51.406525623 +0000 UTC m=+1073.428511237" watchObservedRunningTime="2025-11-23 07:02:51.40831584 +0000 UTC m=+1073.430301454" Nov 23 07:02:56 crc kubenswrapper[4559]: I1123 07:02:56.166745 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:02:56 crc kubenswrapper[4559]: I1123 07:02:56.167274 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:02:58 crc kubenswrapper[4559]: I1123 07:02:58.368773 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-869db7cd47-ghbfq" Nov 23 07:02:58 crc kubenswrapper[4559]: I1123 07:02:58.411539 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-858b54d4df-vf9p9"] Nov 23 07:02:58 crc kubenswrapper[4559]: I1123 07:02:58.412026 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" podUID="92077d58-8565-447b-839d-f830613c4f99" containerName="dnsmasq-dns" containerID="cri-o://26cfd11e9113e63055219afc73e75d789ebbf82df81a54d2363d4b223a062e83" gracePeriod=10 Nov 23 07:02:58 crc kubenswrapper[4559]: I1123 07:02:58.807264 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" Nov 23 07:02:58 crc kubenswrapper[4559]: I1123 07:02:58.876181 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-ovsdbserver-nb\") pod \"92077d58-8565-447b-839d-f830613c4f99\" (UID: \"92077d58-8565-447b-839d-f830613c4f99\") " Nov 23 07:02:58 crc kubenswrapper[4559]: I1123 07:02:58.876268 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-dns-swift-storage-0\") pod \"92077d58-8565-447b-839d-f830613c4f99\" (UID: \"92077d58-8565-447b-839d-f830613c4f99\") " Nov 23 07:02:58 crc kubenswrapper[4559]: I1123 07:02:58.876336 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-config\") pod \"92077d58-8565-447b-839d-f830613c4f99\" (UID: \"92077d58-8565-447b-839d-f830613c4f99\") " Nov 23 07:02:58 crc kubenswrapper[4559]: I1123 07:02:58.876384 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgt9d\" (UniqueName: \"kubernetes.io/projected/92077d58-8565-447b-839d-f830613c4f99-kube-api-access-jgt9d\") pod \"92077d58-8565-447b-839d-f830613c4f99\" (UID: \"92077d58-8565-447b-839d-f830613c4f99\") " Nov 23 07:02:58 crc kubenswrapper[4559]: I1123 07:02:58.876936 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-ovsdbserver-sb\") pod \"92077d58-8565-447b-839d-f830613c4f99\" (UID: \"92077d58-8565-447b-839d-f830613c4f99\") " Nov 23 07:02:58 crc kubenswrapper[4559]: I1123 07:02:58.876991 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-dns-svc\") pod \"92077d58-8565-447b-839d-f830613c4f99\" (UID: \"92077d58-8565-447b-839d-f830613c4f99\") " Nov 23 07:02:58 crc kubenswrapper[4559]: I1123 07:02:58.881089 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92077d58-8565-447b-839d-f830613c4f99-kube-api-access-jgt9d" (OuterVolumeSpecName: "kube-api-access-jgt9d") pod "92077d58-8565-447b-839d-f830613c4f99" (UID: "92077d58-8565-447b-839d-f830613c4f99"). InnerVolumeSpecName "kube-api-access-jgt9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:02:58 crc kubenswrapper[4559]: I1123 07:02:58.912758 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-config" (OuterVolumeSpecName: "config") pod "92077d58-8565-447b-839d-f830613c4f99" (UID: "92077d58-8565-447b-839d-f830613c4f99"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:02:58 crc kubenswrapper[4559]: I1123 07:02:58.916214 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "92077d58-8565-447b-839d-f830613c4f99" (UID: "92077d58-8565-447b-839d-f830613c4f99"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:02:58 crc kubenswrapper[4559]: I1123 07:02:58.920322 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "92077d58-8565-447b-839d-f830613c4f99" (UID: "92077d58-8565-447b-839d-f830613c4f99"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:02:58 crc kubenswrapper[4559]: I1123 07:02:58.922193 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "92077d58-8565-447b-839d-f830613c4f99" (UID: "92077d58-8565-447b-839d-f830613c4f99"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:02:58 crc kubenswrapper[4559]: I1123 07:02:58.922517 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "92077d58-8565-447b-839d-f830613c4f99" (UID: "92077d58-8565-447b-839d-f830613c4f99"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:02:58 crc kubenswrapper[4559]: I1123 07:02:58.979341 4559 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:58 crc kubenswrapper[4559]: I1123 07:02:58.979365 4559 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:58 crc kubenswrapper[4559]: I1123 07:02:58.979374 4559 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:58 crc kubenswrapper[4559]: I1123 07:02:58.979382 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgt9d\" (UniqueName: \"kubernetes.io/projected/92077d58-8565-447b-839d-f830613c4f99-kube-api-access-jgt9d\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:58 crc kubenswrapper[4559]: I1123 07:02:58.979391 4559 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:58 crc kubenswrapper[4559]: I1123 07:02:58.979398 4559 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92077d58-8565-447b-839d-f830613c4f99-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 07:02:59 crc kubenswrapper[4559]: I1123 07:02:59.474062 4559 generic.go:334] "Generic (PLEG): container finished" podID="92077d58-8565-447b-839d-f830613c4f99" containerID="26cfd11e9113e63055219afc73e75d789ebbf82df81a54d2363d4b223a062e83" exitCode=0 Nov 23 07:02:59 crc kubenswrapper[4559]: I1123 07:02:59.474110 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" event={"ID":"92077d58-8565-447b-839d-f830613c4f99","Type":"ContainerDied","Data":"26cfd11e9113e63055219afc73e75d789ebbf82df81a54d2363d4b223a062e83"} Nov 23 07:02:59 crc kubenswrapper[4559]: I1123 07:02:59.474115 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" Nov 23 07:02:59 crc kubenswrapper[4559]: I1123 07:02:59.474143 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-858b54d4df-vf9p9" event={"ID":"92077d58-8565-447b-839d-f830613c4f99","Type":"ContainerDied","Data":"b3d91a0cc7169fd6a3a85da207e9da9fc523ff99c70c2f1f3909d1cf27293ae8"} Nov 23 07:02:59 crc kubenswrapper[4559]: I1123 07:02:59.474166 4559 scope.go:117] "RemoveContainer" containerID="26cfd11e9113e63055219afc73e75d789ebbf82df81a54d2363d4b223a062e83" Nov 23 07:02:59 crc kubenswrapper[4559]: I1123 07:02:59.494891 4559 scope.go:117] "RemoveContainer" containerID="917a01baee638d05e9c0544a5b67da3dd33d8892fbc2c266bffb1f7352ffff34" Nov 23 07:02:59 crc kubenswrapper[4559]: I1123 07:02:59.500107 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-858b54d4df-vf9p9"] Nov 23 07:02:59 crc kubenswrapper[4559]: I1123 07:02:59.505441 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-858b54d4df-vf9p9"] Nov 23 07:02:59 crc kubenswrapper[4559]: I1123 07:02:59.532298 4559 scope.go:117] "RemoveContainer" containerID="26cfd11e9113e63055219afc73e75d789ebbf82df81a54d2363d4b223a062e83" Nov 23 07:02:59 crc kubenswrapper[4559]: E1123 07:02:59.532891 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26cfd11e9113e63055219afc73e75d789ebbf82df81a54d2363d4b223a062e83\": container with ID starting with 26cfd11e9113e63055219afc73e75d789ebbf82df81a54d2363d4b223a062e83 not found: ID does not exist" containerID="26cfd11e9113e63055219afc73e75d789ebbf82df81a54d2363d4b223a062e83" Nov 23 07:02:59 crc kubenswrapper[4559]: I1123 07:02:59.532926 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26cfd11e9113e63055219afc73e75d789ebbf82df81a54d2363d4b223a062e83"} err="failed to get container status \"26cfd11e9113e63055219afc73e75d789ebbf82df81a54d2363d4b223a062e83\": rpc error: code = NotFound desc = could not find container \"26cfd11e9113e63055219afc73e75d789ebbf82df81a54d2363d4b223a062e83\": container with ID starting with 26cfd11e9113e63055219afc73e75d789ebbf82df81a54d2363d4b223a062e83 not found: ID does not exist" Nov 23 07:02:59 crc kubenswrapper[4559]: I1123 07:02:59.532949 4559 scope.go:117] "RemoveContainer" containerID="917a01baee638d05e9c0544a5b67da3dd33d8892fbc2c266bffb1f7352ffff34" Nov 23 07:02:59 crc kubenswrapper[4559]: E1123 07:02:59.533611 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"917a01baee638d05e9c0544a5b67da3dd33d8892fbc2c266bffb1f7352ffff34\": container with ID starting with 917a01baee638d05e9c0544a5b67da3dd33d8892fbc2c266bffb1f7352ffff34 not found: ID does not exist" containerID="917a01baee638d05e9c0544a5b67da3dd33d8892fbc2c266bffb1f7352ffff34" Nov 23 07:02:59 crc kubenswrapper[4559]: I1123 07:02:59.533634 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"917a01baee638d05e9c0544a5b67da3dd33d8892fbc2c266bffb1f7352ffff34"} err="failed to get container status \"917a01baee638d05e9c0544a5b67da3dd33d8892fbc2c266bffb1f7352ffff34\": rpc error: code = NotFound desc = could not find container \"917a01baee638d05e9c0544a5b67da3dd33d8892fbc2c266bffb1f7352ffff34\": container with ID starting with 917a01baee638d05e9c0544a5b67da3dd33d8892fbc2c266bffb1f7352ffff34 not found: ID does not exist" Nov 23 07:03:00 crc kubenswrapper[4559]: I1123 07:03:00.281770 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92077d58-8565-447b-839d-f830613c4f99" path="/var/lib/kubelet/pods/92077d58-8565-447b-839d-f830613c4f99/volumes" Nov 23 07:03:10 crc kubenswrapper[4559]: I1123 07:03:10.310276 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb"] Nov 23 07:03:10 crc kubenswrapper[4559]: E1123 07:03:10.311264 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92077d58-8565-447b-839d-f830613c4f99" containerName="dnsmasq-dns" Nov 23 07:03:10 crc kubenswrapper[4559]: I1123 07:03:10.311280 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="92077d58-8565-447b-839d-f830613c4f99" containerName="dnsmasq-dns" Nov 23 07:03:10 crc kubenswrapper[4559]: E1123 07:03:10.311302 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92077d58-8565-447b-839d-f830613c4f99" containerName="init" Nov 23 07:03:10 crc kubenswrapper[4559]: I1123 07:03:10.311309 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="92077d58-8565-447b-839d-f830613c4f99" containerName="init" Nov 23 07:03:10 crc kubenswrapper[4559]: I1123 07:03:10.311533 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="92077d58-8565-447b-839d-f830613c4f99" containerName="dnsmasq-dns" Nov 23 07:03:10 crc kubenswrapper[4559]: I1123 07:03:10.312236 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb" Nov 23 07:03:10 crc kubenswrapper[4559]: I1123 07:03:10.314175 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 07:03:10 crc kubenswrapper[4559]: I1123 07:03:10.314204 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 07:03:10 crc kubenswrapper[4559]: I1123 07:03:10.314409 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mj6c" Nov 23 07:03:10 crc kubenswrapper[4559]: I1123 07:03:10.314420 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 07:03:10 crc kubenswrapper[4559]: I1123 07:03:10.322518 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb"] Nov 23 07:03:10 crc kubenswrapper[4559]: I1123 07:03:10.477579 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5fa1816-8789-4600-931c-8a376f0936c0-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb\" (UID: \"c5fa1816-8789-4600-931c-8a376f0936c0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb" Nov 23 07:03:10 crc kubenswrapper[4559]: I1123 07:03:10.477776 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5fa1816-8789-4600-931c-8a376f0936c0-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb\" (UID: \"c5fa1816-8789-4600-931c-8a376f0936c0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb" Nov 23 07:03:10 crc kubenswrapper[4559]: I1123 07:03:10.478041 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s44tm\" (UniqueName: \"kubernetes.io/projected/c5fa1816-8789-4600-931c-8a376f0936c0-kube-api-access-s44tm\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb\" (UID: \"c5fa1816-8789-4600-931c-8a376f0936c0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb" Nov 23 07:03:10 crc kubenswrapper[4559]: I1123 07:03:10.478117 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5fa1816-8789-4600-931c-8a376f0936c0-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb\" (UID: \"c5fa1816-8789-4600-931c-8a376f0936c0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb" Nov 23 07:03:10 crc kubenswrapper[4559]: I1123 07:03:10.580686 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5fa1816-8789-4600-931c-8a376f0936c0-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb\" (UID: \"c5fa1816-8789-4600-931c-8a376f0936c0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb" Nov 23 07:03:10 crc kubenswrapper[4559]: I1123 07:03:10.580775 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5fa1816-8789-4600-931c-8a376f0936c0-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb\" (UID: \"c5fa1816-8789-4600-931c-8a376f0936c0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb" Nov 23 07:03:10 crc kubenswrapper[4559]: I1123 07:03:10.580895 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s44tm\" (UniqueName: \"kubernetes.io/projected/c5fa1816-8789-4600-931c-8a376f0936c0-kube-api-access-s44tm\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb\" (UID: \"c5fa1816-8789-4600-931c-8a376f0936c0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb" Nov 23 07:03:10 crc kubenswrapper[4559]: I1123 07:03:10.580945 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5fa1816-8789-4600-931c-8a376f0936c0-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb\" (UID: \"c5fa1816-8789-4600-931c-8a376f0936c0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb" Nov 23 07:03:10 crc kubenswrapper[4559]: I1123 07:03:10.586880 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5fa1816-8789-4600-931c-8a376f0936c0-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb\" (UID: \"c5fa1816-8789-4600-931c-8a376f0936c0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb" Nov 23 07:03:10 crc kubenswrapper[4559]: I1123 07:03:10.587507 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5fa1816-8789-4600-931c-8a376f0936c0-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb\" (UID: \"c5fa1816-8789-4600-931c-8a376f0936c0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb" Nov 23 07:03:10 crc kubenswrapper[4559]: I1123 07:03:10.588804 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5fa1816-8789-4600-931c-8a376f0936c0-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb\" (UID: \"c5fa1816-8789-4600-931c-8a376f0936c0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb" Nov 23 07:03:10 crc kubenswrapper[4559]: I1123 07:03:10.596987 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s44tm\" (UniqueName: \"kubernetes.io/projected/c5fa1816-8789-4600-931c-8a376f0936c0-kube-api-access-s44tm\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb\" (UID: \"c5fa1816-8789-4600-931c-8a376f0936c0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb" Nov 23 07:03:10 crc kubenswrapper[4559]: I1123 07:03:10.630279 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb" Nov 23 07:03:11 crc kubenswrapper[4559]: I1123 07:03:11.088408 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb"] Nov 23 07:03:11 crc kubenswrapper[4559]: W1123 07:03:11.091619 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5fa1816_8789_4600_931c_8a376f0936c0.slice/crio-aeb5e234d8fbe70d98a4d7b0849ce39ce2cf781dc02eed26c6a4e44cc1842261 WatchSource:0}: Error finding container aeb5e234d8fbe70d98a4d7b0849ce39ce2cf781dc02eed26c6a4e44cc1842261: Status 404 returned error can't find the container with id aeb5e234d8fbe70d98a4d7b0849ce39ce2cf781dc02eed26c6a4e44cc1842261 Nov 23 07:03:11 crc kubenswrapper[4559]: I1123 07:03:11.566259 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb" event={"ID":"c5fa1816-8789-4600-931c-8a376f0936c0","Type":"ContainerStarted","Data":"aeb5e234d8fbe70d98a4d7b0849ce39ce2cf781dc02eed26c6a4e44cc1842261"} Nov 23 07:03:19 crc kubenswrapper[4559]: I1123 07:03:19.653053 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb" event={"ID":"c5fa1816-8789-4600-931c-8a376f0936c0","Type":"ContainerStarted","Data":"8d58b28a01fd17b494586bdb2c106f49cd4994ec6aad199a27e5013234c07751"} Nov 23 07:03:19 crc kubenswrapper[4559]: I1123 07:03:19.672103 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb" podStartSLOduration=2.173505888 podStartE2EDuration="9.672075748s" podCreationTimestamp="2025-11-23 07:03:10 +0000 UTC" firstStartedPulling="2025-11-23 07:03:11.093967615 +0000 UTC m=+1093.115953229" lastFinishedPulling="2025-11-23 07:03:18.592537475 +0000 UTC m=+1100.614523089" observedRunningTime="2025-11-23 07:03:19.667877723 +0000 UTC m=+1101.689863337" watchObservedRunningTime="2025-11-23 07:03:19.672075748 +0000 UTC m=+1101.694061363" Nov 23 07:03:20 crc kubenswrapper[4559]: I1123 07:03:20.669020 4559 generic.go:334] "Generic (PLEG): container finished" podID="e0731be9-7300-45df-b028-8aeb2a3999b4" containerID="60cd43dfdb5398753bd255ddfb13142a17a3642c8d822ebc86fdda896d0d281a" exitCode=0 Nov 23 07:03:20 crc kubenswrapper[4559]: I1123 07:03:20.669119 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e0731be9-7300-45df-b028-8aeb2a3999b4","Type":"ContainerDied","Data":"60cd43dfdb5398753bd255ddfb13142a17a3642c8d822ebc86fdda896d0d281a"} Nov 23 07:03:20 crc kubenswrapper[4559]: I1123 07:03:20.671737 4559 generic.go:334] "Generic (PLEG): container finished" podID="bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8" containerID="2b13ac54b4bcbbd498556afe316007d3125b4fcde79e9bfef0769f0d7dff618e" exitCode=0 Nov 23 07:03:20 crc kubenswrapper[4559]: I1123 07:03:20.671806 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8","Type":"ContainerDied","Data":"2b13ac54b4bcbbd498556afe316007d3125b4fcde79e9bfef0769f0d7dff618e"} Nov 23 07:03:21 crc kubenswrapper[4559]: I1123 07:03:21.694839 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8","Type":"ContainerStarted","Data":"e947b6d56e19b48dfd2c00b88af3ae6e5a4143e4484e81fb4c4cf4c4b8788f92"} Nov 23 07:03:21 crc kubenswrapper[4559]: I1123 07:03:21.695709 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 23 07:03:21 crc kubenswrapper[4559]: I1123 07:03:21.698974 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e0731be9-7300-45df-b028-8aeb2a3999b4","Type":"ContainerStarted","Data":"ad624bcf1c2b5f6b32389e300f37ecbb0802e05161f5b01fc995c9859e7e76c1"} Nov 23 07:03:21 crc kubenswrapper[4559]: I1123 07:03:21.699245 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:03:21 crc kubenswrapper[4559]: I1123 07:03:21.717416 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=35.717396693 podStartE2EDuration="35.717396693s" podCreationTimestamp="2025-11-23 07:02:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:03:21.71191783 +0000 UTC m=+1103.733903444" watchObservedRunningTime="2025-11-23 07:03:21.717396693 +0000 UTC m=+1103.739382307" Nov 23 07:03:21 crc kubenswrapper[4559]: I1123 07:03:21.735602 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=34.735575141 podStartE2EDuration="34.735575141s" podCreationTimestamp="2025-11-23 07:02:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:03:21.728179011 +0000 UTC m=+1103.750164625" watchObservedRunningTime="2025-11-23 07:03:21.735575141 +0000 UTC m=+1103.757560754" Nov 23 07:03:26 crc kubenswrapper[4559]: I1123 07:03:26.167272 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:03:26 crc kubenswrapper[4559]: I1123 07:03:26.167745 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:03:26 crc kubenswrapper[4559]: I1123 07:03:26.167784 4559 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" Nov 23 07:03:26 crc kubenswrapper[4559]: I1123 07:03:26.168410 4559 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b7be9dbebcbc9b258e6fc583ce1dee8a3804425c02033654afb5190f533977ae"} pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 07:03:26 crc kubenswrapper[4559]: I1123 07:03:26.168460 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" containerID="cri-o://b7be9dbebcbc9b258e6fc583ce1dee8a3804425c02033654afb5190f533977ae" gracePeriod=600 Nov 23 07:03:26 crc kubenswrapper[4559]: I1123 07:03:26.749460 4559 generic.go:334] "Generic (PLEG): container finished" podID="4731beee-0cac-4189-8a70-743b0b709095" containerID="b7be9dbebcbc9b258e6fc583ce1dee8a3804425c02033654afb5190f533977ae" exitCode=0 Nov 23 07:03:26 crc kubenswrapper[4559]: I1123 07:03:26.749500 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerDied","Data":"b7be9dbebcbc9b258e6fc583ce1dee8a3804425c02033654afb5190f533977ae"} Nov 23 07:03:26 crc kubenswrapper[4559]: I1123 07:03:26.749978 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerStarted","Data":"7f8ad51d7a50444a4cc31965b149316d371db5ed9346214260de65924e42fa55"} Nov 23 07:03:26 crc kubenswrapper[4559]: I1123 07:03:26.750016 4559 scope.go:117] "RemoveContainer" containerID="667a247ddea953ad2f29e458b8a21398c7c8c20545d619492b9346eff264de8f" Nov 23 07:03:30 crc kubenswrapper[4559]: I1123 07:03:30.804512 4559 generic.go:334] "Generic (PLEG): container finished" podID="c5fa1816-8789-4600-931c-8a376f0936c0" containerID="8d58b28a01fd17b494586bdb2c106f49cd4994ec6aad199a27e5013234c07751" exitCode=0 Nov 23 07:03:30 crc kubenswrapper[4559]: I1123 07:03:30.804874 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb" event={"ID":"c5fa1816-8789-4600-931c-8a376f0936c0","Type":"ContainerDied","Data":"8d58b28a01fd17b494586bdb2c106f49cd4994ec6aad199a27e5013234c07751"} Nov 23 07:03:32 crc kubenswrapper[4559]: I1123 07:03:32.109209 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb" Nov 23 07:03:32 crc kubenswrapper[4559]: I1123 07:03:32.252030 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5fa1816-8789-4600-931c-8a376f0936c0-ssh-key\") pod \"c5fa1816-8789-4600-931c-8a376f0936c0\" (UID: \"c5fa1816-8789-4600-931c-8a376f0936c0\") " Nov 23 07:03:32 crc kubenswrapper[4559]: I1123 07:03:32.252186 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s44tm\" (UniqueName: \"kubernetes.io/projected/c5fa1816-8789-4600-931c-8a376f0936c0-kube-api-access-s44tm\") pod \"c5fa1816-8789-4600-931c-8a376f0936c0\" (UID: \"c5fa1816-8789-4600-931c-8a376f0936c0\") " Nov 23 07:03:32 crc kubenswrapper[4559]: I1123 07:03:32.252270 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5fa1816-8789-4600-931c-8a376f0936c0-repo-setup-combined-ca-bundle\") pod \"c5fa1816-8789-4600-931c-8a376f0936c0\" (UID: \"c5fa1816-8789-4600-931c-8a376f0936c0\") " Nov 23 07:03:32 crc kubenswrapper[4559]: I1123 07:03:32.252372 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5fa1816-8789-4600-931c-8a376f0936c0-inventory\") pod \"c5fa1816-8789-4600-931c-8a376f0936c0\" (UID: \"c5fa1816-8789-4600-931c-8a376f0936c0\") " Nov 23 07:03:32 crc kubenswrapper[4559]: I1123 07:03:32.257791 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5fa1816-8789-4600-931c-8a376f0936c0-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "c5fa1816-8789-4600-931c-8a376f0936c0" (UID: "c5fa1816-8789-4600-931c-8a376f0936c0"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:03:32 crc kubenswrapper[4559]: I1123 07:03:32.258187 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5fa1816-8789-4600-931c-8a376f0936c0-kube-api-access-s44tm" (OuterVolumeSpecName: "kube-api-access-s44tm") pod "c5fa1816-8789-4600-931c-8a376f0936c0" (UID: "c5fa1816-8789-4600-931c-8a376f0936c0"). InnerVolumeSpecName "kube-api-access-s44tm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:03:32 crc kubenswrapper[4559]: I1123 07:03:32.273742 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5fa1816-8789-4600-931c-8a376f0936c0-inventory" (OuterVolumeSpecName: "inventory") pod "c5fa1816-8789-4600-931c-8a376f0936c0" (UID: "c5fa1816-8789-4600-931c-8a376f0936c0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:03:32 crc kubenswrapper[4559]: I1123 07:03:32.274014 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5fa1816-8789-4600-931c-8a376f0936c0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c5fa1816-8789-4600-931c-8a376f0936c0" (UID: "c5fa1816-8789-4600-931c-8a376f0936c0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:03:32 crc kubenswrapper[4559]: I1123 07:03:32.355157 4559 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5fa1816-8789-4600-931c-8a376f0936c0-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:03:32 crc kubenswrapper[4559]: I1123 07:03:32.355187 4559 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5fa1816-8789-4600-931c-8a376f0936c0-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 07:03:32 crc kubenswrapper[4559]: I1123 07:03:32.355479 4559 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5fa1816-8789-4600-931c-8a376f0936c0-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 07:03:32 crc kubenswrapper[4559]: I1123 07:03:32.355490 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s44tm\" (UniqueName: \"kubernetes.io/projected/c5fa1816-8789-4600-931c-8a376f0936c0-kube-api-access-s44tm\") on node \"crc\" DevicePath \"\"" Nov 23 07:03:32 crc kubenswrapper[4559]: I1123 07:03:32.821090 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb" event={"ID":"c5fa1816-8789-4600-931c-8a376f0936c0","Type":"ContainerDied","Data":"aeb5e234d8fbe70d98a4d7b0849ce39ce2cf781dc02eed26c6a4e44cc1842261"} Nov 23 07:03:32 crc kubenswrapper[4559]: I1123 07:03:32.821138 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aeb5e234d8fbe70d98a4d7b0849ce39ce2cf781dc02eed26c6a4e44cc1842261" Nov 23 07:03:32 crc kubenswrapper[4559]: I1123 07:03:32.821167 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb" Nov 23 07:03:32 crc kubenswrapper[4559]: I1123 07:03:32.866451 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-sx224"] Nov 23 07:03:32 crc kubenswrapper[4559]: E1123 07:03:32.867174 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5fa1816-8789-4600-931c-8a376f0936c0" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 23 07:03:32 crc kubenswrapper[4559]: I1123 07:03:32.867343 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5fa1816-8789-4600-931c-8a376f0936c0" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 23 07:03:32 crc kubenswrapper[4559]: I1123 07:03:32.867616 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5fa1816-8789-4600-931c-8a376f0936c0" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 23 07:03:32 crc kubenswrapper[4559]: I1123 07:03:32.868282 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-sx224" Nov 23 07:03:32 crc kubenswrapper[4559]: I1123 07:03:32.870705 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 07:03:32 crc kubenswrapper[4559]: I1123 07:03:32.870794 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 07:03:32 crc kubenswrapper[4559]: I1123 07:03:32.871183 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mj6c" Nov 23 07:03:32 crc kubenswrapper[4559]: I1123 07:03:32.871257 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 07:03:32 crc kubenswrapper[4559]: I1123 07:03:32.879577 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-sx224"] Nov 23 07:03:32 crc kubenswrapper[4559]: I1123 07:03:32.966510 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b84c748-5838-4938-be55-0e8fc4174730-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-sx224\" (UID: \"1b84c748-5838-4938-be55-0e8fc4174730\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-sx224" Nov 23 07:03:32 crc kubenswrapper[4559]: I1123 07:03:32.966684 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6mmw\" (UniqueName: \"kubernetes.io/projected/1b84c748-5838-4938-be55-0e8fc4174730-kube-api-access-b6mmw\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-sx224\" (UID: \"1b84c748-5838-4938-be55-0e8fc4174730\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-sx224" Nov 23 07:03:32 crc kubenswrapper[4559]: I1123 07:03:32.966790 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b84c748-5838-4938-be55-0e8fc4174730-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-sx224\" (UID: \"1b84c748-5838-4938-be55-0e8fc4174730\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-sx224" Nov 23 07:03:33 crc kubenswrapper[4559]: I1123 07:03:33.068445 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6mmw\" (UniqueName: \"kubernetes.io/projected/1b84c748-5838-4938-be55-0e8fc4174730-kube-api-access-b6mmw\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-sx224\" (UID: \"1b84c748-5838-4938-be55-0e8fc4174730\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-sx224" Nov 23 07:03:33 crc kubenswrapper[4559]: I1123 07:03:33.068502 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b84c748-5838-4938-be55-0e8fc4174730-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-sx224\" (UID: \"1b84c748-5838-4938-be55-0e8fc4174730\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-sx224" Nov 23 07:03:33 crc kubenswrapper[4559]: I1123 07:03:33.068690 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b84c748-5838-4938-be55-0e8fc4174730-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-sx224\" (UID: \"1b84c748-5838-4938-be55-0e8fc4174730\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-sx224" Nov 23 07:03:33 crc kubenswrapper[4559]: I1123 07:03:33.074099 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b84c748-5838-4938-be55-0e8fc4174730-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-sx224\" (UID: \"1b84c748-5838-4938-be55-0e8fc4174730\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-sx224" Nov 23 07:03:33 crc kubenswrapper[4559]: I1123 07:03:33.074510 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b84c748-5838-4938-be55-0e8fc4174730-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-sx224\" (UID: \"1b84c748-5838-4938-be55-0e8fc4174730\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-sx224" Nov 23 07:03:33 crc kubenswrapper[4559]: I1123 07:03:33.081382 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6mmw\" (UniqueName: \"kubernetes.io/projected/1b84c748-5838-4938-be55-0e8fc4174730-kube-api-access-b6mmw\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-sx224\" (UID: \"1b84c748-5838-4938-be55-0e8fc4174730\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-sx224" Nov 23 07:03:33 crc kubenswrapper[4559]: I1123 07:03:33.190615 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-sx224" Nov 23 07:03:33 crc kubenswrapper[4559]: I1123 07:03:33.608992 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-sx224"] Nov 23 07:03:33 crc kubenswrapper[4559]: W1123 07:03:33.617773 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b84c748_5838_4938_be55_0e8fc4174730.slice/crio-9421e6afb22ac060611c2688324bcc5025d89baa3dd22c6139f1742babf840e6 WatchSource:0}: Error finding container 9421e6afb22ac060611c2688324bcc5025d89baa3dd22c6139f1742babf840e6: Status 404 returned error can't find the container with id 9421e6afb22ac060611c2688324bcc5025d89baa3dd22c6139f1742babf840e6 Nov 23 07:03:33 crc kubenswrapper[4559]: I1123 07:03:33.830839 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-sx224" event={"ID":"1b84c748-5838-4938-be55-0e8fc4174730","Type":"ContainerStarted","Data":"9421e6afb22ac060611c2688324bcc5025d89baa3dd22c6139f1742babf840e6"} Nov 23 07:03:34 crc kubenswrapper[4559]: I1123 07:03:34.838604 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-sx224" event={"ID":"1b84c748-5838-4938-be55-0e8fc4174730","Type":"ContainerStarted","Data":"3a42994c71b1108a67e7d2f2cb74a13e25a41677718750d69b3f80cf124e1192"} Nov 23 07:03:34 crc kubenswrapper[4559]: I1123 07:03:34.856063 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-sx224" podStartSLOduration=2.383876126 podStartE2EDuration="2.856045884s" podCreationTimestamp="2025-11-23 07:03:32 +0000 UTC" firstStartedPulling="2025-11-23 07:03:33.620830759 +0000 UTC m=+1115.642816373" lastFinishedPulling="2025-11-23 07:03:34.093000517 +0000 UTC m=+1116.114986131" observedRunningTime="2025-11-23 07:03:34.853185915 +0000 UTC m=+1116.875171530" watchObservedRunningTime="2025-11-23 07:03:34.856045884 +0000 UTC m=+1116.878031499" Nov 23 07:03:36 crc kubenswrapper[4559]: I1123 07:03:36.709842 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 23 07:03:36 crc kubenswrapper[4559]: I1123 07:03:36.856555 4559 generic.go:334] "Generic (PLEG): container finished" podID="1b84c748-5838-4938-be55-0e8fc4174730" containerID="3a42994c71b1108a67e7d2f2cb74a13e25a41677718750d69b3f80cf124e1192" exitCode=0 Nov 23 07:03:36 crc kubenswrapper[4559]: I1123 07:03:36.856605 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-sx224" event={"ID":"1b84c748-5838-4938-be55-0e8fc4174730","Type":"ContainerDied","Data":"3a42994c71b1108a67e7d2f2cb74a13e25a41677718750d69b3f80cf124e1192"} Nov 23 07:03:37 crc kubenswrapper[4559]: I1123 07:03:37.737806 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 23 07:03:38 crc kubenswrapper[4559]: I1123 07:03:38.198135 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-sx224" Nov 23 07:03:38 crc kubenswrapper[4559]: I1123 07:03:38.266252 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6mmw\" (UniqueName: \"kubernetes.io/projected/1b84c748-5838-4938-be55-0e8fc4174730-kube-api-access-b6mmw\") pod \"1b84c748-5838-4938-be55-0e8fc4174730\" (UID: \"1b84c748-5838-4938-be55-0e8fc4174730\") " Nov 23 07:03:38 crc kubenswrapper[4559]: I1123 07:03:38.266393 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b84c748-5838-4938-be55-0e8fc4174730-ssh-key\") pod \"1b84c748-5838-4938-be55-0e8fc4174730\" (UID: \"1b84c748-5838-4938-be55-0e8fc4174730\") " Nov 23 07:03:38 crc kubenswrapper[4559]: I1123 07:03:38.266426 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b84c748-5838-4938-be55-0e8fc4174730-inventory\") pod \"1b84c748-5838-4938-be55-0e8fc4174730\" (UID: \"1b84c748-5838-4938-be55-0e8fc4174730\") " Nov 23 07:03:38 crc kubenswrapper[4559]: I1123 07:03:38.271815 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b84c748-5838-4938-be55-0e8fc4174730-kube-api-access-b6mmw" (OuterVolumeSpecName: "kube-api-access-b6mmw") pod "1b84c748-5838-4938-be55-0e8fc4174730" (UID: "1b84c748-5838-4938-be55-0e8fc4174730"). InnerVolumeSpecName "kube-api-access-b6mmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:03:38 crc kubenswrapper[4559]: I1123 07:03:38.288595 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b84c748-5838-4938-be55-0e8fc4174730-inventory" (OuterVolumeSpecName: "inventory") pod "1b84c748-5838-4938-be55-0e8fc4174730" (UID: "1b84c748-5838-4938-be55-0e8fc4174730"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:03:38 crc kubenswrapper[4559]: I1123 07:03:38.289676 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b84c748-5838-4938-be55-0e8fc4174730-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1b84c748-5838-4938-be55-0e8fc4174730" (UID: "1b84c748-5838-4938-be55-0e8fc4174730"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:03:38 crc kubenswrapper[4559]: I1123 07:03:38.368450 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6mmw\" (UniqueName: \"kubernetes.io/projected/1b84c748-5838-4938-be55-0e8fc4174730-kube-api-access-b6mmw\") on node \"crc\" DevicePath \"\"" Nov 23 07:03:38 crc kubenswrapper[4559]: I1123 07:03:38.368481 4559 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b84c748-5838-4938-be55-0e8fc4174730-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 07:03:38 crc kubenswrapper[4559]: I1123 07:03:38.368493 4559 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b84c748-5838-4938-be55-0e8fc4174730-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 07:03:38 crc kubenswrapper[4559]: I1123 07:03:38.871102 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-sx224" event={"ID":"1b84c748-5838-4938-be55-0e8fc4174730","Type":"ContainerDied","Data":"9421e6afb22ac060611c2688324bcc5025d89baa3dd22c6139f1742babf840e6"} Nov 23 07:03:38 crc kubenswrapper[4559]: I1123 07:03:38.871738 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9421e6afb22ac060611c2688324bcc5025d89baa3dd22c6139f1742babf840e6" Nov 23 07:03:38 crc kubenswrapper[4559]: I1123 07:03:38.871832 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-sx224" Nov 23 07:03:38 crc kubenswrapper[4559]: I1123 07:03:38.919767 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz"] Nov 23 07:03:38 crc kubenswrapper[4559]: E1123 07:03:38.920234 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b84c748-5838-4938-be55-0e8fc4174730" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 23 07:03:38 crc kubenswrapper[4559]: I1123 07:03:38.920254 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b84c748-5838-4938-be55-0e8fc4174730" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 23 07:03:38 crc kubenswrapper[4559]: I1123 07:03:38.920450 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b84c748-5838-4938-be55-0e8fc4174730" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 23 07:03:38 crc kubenswrapper[4559]: I1123 07:03:38.921109 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz" Nov 23 07:03:38 crc kubenswrapper[4559]: I1123 07:03:38.922577 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 07:03:38 crc kubenswrapper[4559]: I1123 07:03:38.923376 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 07:03:38 crc kubenswrapper[4559]: I1123 07:03:38.923434 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mj6c" Nov 23 07:03:38 crc kubenswrapper[4559]: I1123 07:03:38.924189 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 07:03:38 crc kubenswrapper[4559]: I1123 07:03:38.928246 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz"] Nov 23 07:03:38 crc kubenswrapper[4559]: I1123 07:03:38.979069 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af59bdf5-b08e-487d-bcc4-c70a275cace5-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz\" (UID: \"af59bdf5-b08e-487d-bcc4-c70a275cace5\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz" Nov 23 07:03:38 crc kubenswrapper[4559]: I1123 07:03:38.979169 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af59bdf5-b08e-487d-bcc4-c70a275cace5-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz\" (UID: \"af59bdf5-b08e-487d-bcc4-c70a275cace5\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz" Nov 23 07:03:38 crc kubenswrapper[4559]: I1123 07:03:38.979234 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-449dm\" (UniqueName: \"kubernetes.io/projected/af59bdf5-b08e-487d-bcc4-c70a275cace5-kube-api-access-449dm\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz\" (UID: \"af59bdf5-b08e-487d-bcc4-c70a275cace5\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz" Nov 23 07:03:38 crc kubenswrapper[4559]: I1123 07:03:38.979347 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af59bdf5-b08e-487d-bcc4-c70a275cace5-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz\" (UID: \"af59bdf5-b08e-487d-bcc4-c70a275cace5\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz" Nov 23 07:03:39 crc kubenswrapper[4559]: I1123 07:03:39.080134 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af59bdf5-b08e-487d-bcc4-c70a275cace5-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz\" (UID: \"af59bdf5-b08e-487d-bcc4-c70a275cace5\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz" Nov 23 07:03:39 crc kubenswrapper[4559]: I1123 07:03:39.080218 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af59bdf5-b08e-487d-bcc4-c70a275cace5-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz\" (UID: \"af59bdf5-b08e-487d-bcc4-c70a275cace5\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz" Nov 23 07:03:39 crc kubenswrapper[4559]: I1123 07:03:39.080271 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af59bdf5-b08e-487d-bcc4-c70a275cace5-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz\" (UID: \"af59bdf5-b08e-487d-bcc4-c70a275cace5\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz" Nov 23 07:03:39 crc kubenswrapper[4559]: I1123 07:03:39.080316 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-449dm\" (UniqueName: \"kubernetes.io/projected/af59bdf5-b08e-487d-bcc4-c70a275cace5-kube-api-access-449dm\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz\" (UID: \"af59bdf5-b08e-487d-bcc4-c70a275cace5\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz" Nov 23 07:03:39 crc kubenswrapper[4559]: I1123 07:03:39.084843 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af59bdf5-b08e-487d-bcc4-c70a275cace5-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz\" (UID: \"af59bdf5-b08e-487d-bcc4-c70a275cace5\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz" Nov 23 07:03:39 crc kubenswrapper[4559]: I1123 07:03:39.084961 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af59bdf5-b08e-487d-bcc4-c70a275cace5-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz\" (UID: \"af59bdf5-b08e-487d-bcc4-c70a275cace5\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz" Nov 23 07:03:39 crc kubenswrapper[4559]: I1123 07:03:39.085097 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af59bdf5-b08e-487d-bcc4-c70a275cace5-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz\" (UID: \"af59bdf5-b08e-487d-bcc4-c70a275cace5\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz" Nov 23 07:03:39 crc kubenswrapper[4559]: I1123 07:03:39.094418 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-449dm\" (UniqueName: \"kubernetes.io/projected/af59bdf5-b08e-487d-bcc4-c70a275cace5-kube-api-access-449dm\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz\" (UID: \"af59bdf5-b08e-487d-bcc4-c70a275cace5\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz" Nov 23 07:03:39 crc kubenswrapper[4559]: I1123 07:03:39.235465 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz" Nov 23 07:03:39 crc kubenswrapper[4559]: I1123 07:03:39.676788 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz"] Nov 23 07:03:39 crc kubenswrapper[4559]: W1123 07:03:39.678957 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf59bdf5_b08e_487d_bcc4_c70a275cace5.slice/crio-79bec19b3885e3cc7d667e00d1548d1814110e1611c4739700d9a5655989cf88 WatchSource:0}: Error finding container 79bec19b3885e3cc7d667e00d1548d1814110e1611c4739700d9a5655989cf88: Status 404 returned error can't find the container with id 79bec19b3885e3cc7d667e00d1548d1814110e1611c4739700d9a5655989cf88 Nov 23 07:03:39 crc kubenswrapper[4559]: I1123 07:03:39.681366 4559 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 07:03:39 crc kubenswrapper[4559]: I1123 07:03:39.879030 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz" event={"ID":"af59bdf5-b08e-487d-bcc4-c70a275cace5","Type":"ContainerStarted","Data":"79bec19b3885e3cc7d667e00d1548d1814110e1611c4739700d9a5655989cf88"} Nov 23 07:03:40 crc kubenswrapper[4559]: I1123 07:03:40.886947 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz" event={"ID":"af59bdf5-b08e-487d-bcc4-c70a275cace5","Type":"ContainerStarted","Data":"e6c12d08dca70af35d08a6839e3d13253d492ad4f949f2cfb61f630a0ea763c2"} Nov 23 07:03:40 crc kubenswrapper[4559]: I1123 07:03:40.900233 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz" podStartSLOduration=2.398118286 podStartE2EDuration="2.900214486s" podCreationTimestamp="2025-11-23 07:03:38 +0000 UTC" firstStartedPulling="2025-11-23 07:03:39.681136499 +0000 UTC m=+1121.703122113" lastFinishedPulling="2025-11-23 07:03:40.183232699 +0000 UTC m=+1122.205218313" observedRunningTime="2025-11-23 07:03:40.899531511 +0000 UTC m=+1122.921517125" watchObservedRunningTime="2025-11-23 07:03:40.900214486 +0000 UTC m=+1122.922200100" Nov 23 07:05:12 crc kubenswrapper[4559]: I1123 07:05:12.063466 4559 scope.go:117] "RemoveContainer" containerID="a5990d9b69ba08df670ad9a966377f748277a1de4d043ccfe6259d49aa74c25c" Nov 23 07:05:26 crc kubenswrapper[4559]: I1123 07:05:26.167151 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:05:26 crc kubenswrapper[4559]: I1123 07:05:26.167773 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:05:56 crc kubenswrapper[4559]: I1123 07:05:56.167567 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:05:56 crc kubenswrapper[4559]: I1123 07:05:56.168821 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:06:12 crc kubenswrapper[4559]: I1123 07:06:12.126522 4559 scope.go:117] "RemoveContainer" containerID="8f7af979fa7061247fe8031e3f0c0c907f89b6bdedef6766b14c7cde78cb28fd" Nov 23 07:06:26 crc kubenswrapper[4559]: I1123 07:06:26.166552 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:06:26 crc kubenswrapper[4559]: I1123 07:06:26.166981 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:06:26 crc kubenswrapper[4559]: I1123 07:06:26.167030 4559 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" Nov 23 07:06:26 crc kubenswrapper[4559]: I1123 07:06:26.167847 4559 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7f8ad51d7a50444a4cc31965b149316d371db5ed9346214260de65924e42fa55"} pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 07:06:26 crc kubenswrapper[4559]: I1123 07:06:26.167906 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" containerID="cri-o://7f8ad51d7a50444a4cc31965b149316d371db5ed9346214260de65924e42fa55" gracePeriod=600 Nov 23 07:06:26 crc kubenswrapper[4559]: E1123 07:06:26.350854 4559 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4731beee_0cac_4189_8a70_743b0b709095.slice/crio-conmon-7f8ad51d7a50444a4cc31965b149316d371db5ed9346214260de65924e42fa55.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4731beee_0cac_4189_8a70_743b0b709095.slice/crio-7f8ad51d7a50444a4cc31965b149316d371db5ed9346214260de65924e42fa55.scope\": RecentStats: unable to find data in memory cache]" Nov 23 07:06:27 crc kubenswrapper[4559]: I1123 07:06:27.208874 4559 generic.go:334] "Generic (PLEG): container finished" podID="4731beee-0cac-4189-8a70-743b0b709095" containerID="7f8ad51d7a50444a4cc31965b149316d371db5ed9346214260de65924e42fa55" exitCode=0 Nov 23 07:06:27 crc kubenswrapper[4559]: I1123 07:06:27.208923 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerDied","Data":"7f8ad51d7a50444a4cc31965b149316d371db5ed9346214260de65924e42fa55"} Nov 23 07:06:27 crc kubenswrapper[4559]: I1123 07:06:27.210657 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerStarted","Data":"ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b"} Nov 23 07:06:27 crc kubenswrapper[4559]: I1123 07:06:27.210744 4559 scope.go:117] "RemoveContainer" containerID="b7be9dbebcbc9b258e6fc583ce1dee8a3804425c02033654afb5190f533977ae" Nov 23 07:06:39 crc kubenswrapper[4559]: I1123 07:06:39.320186 4559 generic.go:334] "Generic (PLEG): container finished" podID="af59bdf5-b08e-487d-bcc4-c70a275cace5" containerID="e6c12d08dca70af35d08a6839e3d13253d492ad4f949f2cfb61f630a0ea763c2" exitCode=0 Nov 23 07:06:39 crc kubenswrapper[4559]: I1123 07:06:39.320265 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz" event={"ID":"af59bdf5-b08e-487d-bcc4-c70a275cace5","Type":"ContainerDied","Data":"e6c12d08dca70af35d08a6839e3d13253d492ad4f949f2cfb61f630a0ea763c2"} Nov 23 07:06:40 crc kubenswrapper[4559]: I1123 07:06:40.666705 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz" Nov 23 07:06:40 crc kubenswrapper[4559]: I1123 07:06:40.822159 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-449dm\" (UniqueName: \"kubernetes.io/projected/af59bdf5-b08e-487d-bcc4-c70a275cace5-kube-api-access-449dm\") pod \"af59bdf5-b08e-487d-bcc4-c70a275cace5\" (UID: \"af59bdf5-b08e-487d-bcc4-c70a275cace5\") " Nov 23 07:06:40 crc kubenswrapper[4559]: I1123 07:06:40.822236 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af59bdf5-b08e-487d-bcc4-c70a275cace5-inventory\") pod \"af59bdf5-b08e-487d-bcc4-c70a275cace5\" (UID: \"af59bdf5-b08e-487d-bcc4-c70a275cace5\") " Nov 23 07:06:40 crc kubenswrapper[4559]: I1123 07:06:40.822336 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af59bdf5-b08e-487d-bcc4-c70a275cace5-bootstrap-combined-ca-bundle\") pod \"af59bdf5-b08e-487d-bcc4-c70a275cace5\" (UID: \"af59bdf5-b08e-487d-bcc4-c70a275cace5\") " Nov 23 07:06:40 crc kubenswrapper[4559]: I1123 07:06:40.822427 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af59bdf5-b08e-487d-bcc4-c70a275cace5-ssh-key\") pod \"af59bdf5-b08e-487d-bcc4-c70a275cace5\" (UID: \"af59bdf5-b08e-487d-bcc4-c70a275cace5\") " Nov 23 07:06:40 crc kubenswrapper[4559]: I1123 07:06:40.828412 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af59bdf5-b08e-487d-bcc4-c70a275cace5-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "af59bdf5-b08e-487d-bcc4-c70a275cace5" (UID: "af59bdf5-b08e-487d-bcc4-c70a275cace5"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:06:40 crc kubenswrapper[4559]: I1123 07:06:40.828503 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af59bdf5-b08e-487d-bcc4-c70a275cace5-kube-api-access-449dm" (OuterVolumeSpecName: "kube-api-access-449dm") pod "af59bdf5-b08e-487d-bcc4-c70a275cace5" (UID: "af59bdf5-b08e-487d-bcc4-c70a275cace5"). InnerVolumeSpecName "kube-api-access-449dm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:06:40 crc kubenswrapper[4559]: I1123 07:06:40.845794 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af59bdf5-b08e-487d-bcc4-c70a275cace5-inventory" (OuterVolumeSpecName: "inventory") pod "af59bdf5-b08e-487d-bcc4-c70a275cace5" (UID: "af59bdf5-b08e-487d-bcc4-c70a275cace5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:06:40 crc kubenswrapper[4559]: I1123 07:06:40.847594 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af59bdf5-b08e-487d-bcc4-c70a275cace5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "af59bdf5-b08e-487d-bcc4-c70a275cace5" (UID: "af59bdf5-b08e-487d-bcc4-c70a275cace5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:06:40 crc kubenswrapper[4559]: I1123 07:06:40.925216 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-449dm\" (UniqueName: \"kubernetes.io/projected/af59bdf5-b08e-487d-bcc4-c70a275cace5-kube-api-access-449dm\") on node \"crc\" DevicePath \"\"" Nov 23 07:06:40 crc kubenswrapper[4559]: I1123 07:06:40.925249 4559 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af59bdf5-b08e-487d-bcc4-c70a275cace5-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 07:06:40 crc kubenswrapper[4559]: I1123 07:06:40.925261 4559 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af59bdf5-b08e-487d-bcc4-c70a275cace5-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:06:40 crc kubenswrapper[4559]: I1123 07:06:40.925270 4559 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af59bdf5-b08e-487d-bcc4-c70a275cace5-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 07:06:41 crc kubenswrapper[4559]: I1123 07:06:41.346557 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz" Nov 23 07:06:41 crc kubenswrapper[4559]: I1123 07:06:41.346491 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz" event={"ID":"af59bdf5-b08e-487d-bcc4-c70a275cace5","Type":"ContainerDied","Data":"79bec19b3885e3cc7d667e00d1548d1814110e1611c4739700d9a5655989cf88"} Nov 23 07:06:41 crc kubenswrapper[4559]: I1123 07:06:41.346922 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79bec19b3885e3cc7d667e00d1548d1814110e1611c4739700d9a5655989cf88" Nov 23 07:06:41 crc kubenswrapper[4559]: I1123 07:06:41.402995 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46r2s"] Nov 23 07:06:41 crc kubenswrapper[4559]: E1123 07:06:41.403363 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af59bdf5-b08e-487d-bcc4-c70a275cace5" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 23 07:06:41 crc kubenswrapper[4559]: I1123 07:06:41.403382 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="af59bdf5-b08e-487d-bcc4-c70a275cace5" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 23 07:06:41 crc kubenswrapper[4559]: I1123 07:06:41.403583 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="af59bdf5-b08e-487d-bcc4-c70a275cace5" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 23 07:06:41 crc kubenswrapper[4559]: I1123 07:06:41.404170 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46r2s" Nov 23 07:06:41 crc kubenswrapper[4559]: I1123 07:06:41.408962 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 07:06:41 crc kubenswrapper[4559]: I1123 07:06:41.408962 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mj6c" Nov 23 07:06:41 crc kubenswrapper[4559]: I1123 07:06:41.409017 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 07:06:41 crc kubenswrapper[4559]: I1123 07:06:41.409025 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 07:06:41 crc kubenswrapper[4559]: I1123 07:06:41.410913 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46r2s"] Nov 23 07:06:41 crc kubenswrapper[4559]: I1123 07:06:41.533546 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-st6jn\" (UniqueName: \"kubernetes.io/projected/fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29-kube-api-access-st6jn\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-46r2s\" (UID: \"fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46r2s" Nov 23 07:06:41 crc kubenswrapper[4559]: I1123 07:06:41.534055 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-46r2s\" (UID: \"fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46r2s" Nov 23 07:06:41 crc kubenswrapper[4559]: I1123 07:06:41.534177 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-46r2s\" (UID: \"fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46r2s" Nov 23 07:06:41 crc kubenswrapper[4559]: I1123 07:06:41.636231 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-st6jn\" (UniqueName: \"kubernetes.io/projected/fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29-kube-api-access-st6jn\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-46r2s\" (UID: \"fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46r2s" Nov 23 07:06:41 crc kubenswrapper[4559]: I1123 07:06:41.636448 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-46r2s\" (UID: \"fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46r2s" Nov 23 07:06:41 crc kubenswrapper[4559]: I1123 07:06:41.636523 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-46r2s\" (UID: \"fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46r2s" Nov 23 07:06:41 crc kubenswrapper[4559]: I1123 07:06:41.640066 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-46r2s\" (UID: \"fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46r2s" Nov 23 07:06:41 crc kubenswrapper[4559]: I1123 07:06:41.640075 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-46r2s\" (UID: \"fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46r2s" Nov 23 07:06:41 crc kubenswrapper[4559]: I1123 07:06:41.650275 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-st6jn\" (UniqueName: \"kubernetes.io/projected/fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29-kube-api-access-st6jn\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-46r2s\" (UID: \"fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46r2s" Nov 23 07:06:41 crc kubenswrapper[4559]: I1123 07:06:41.722808 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46r2s" Nov 23 07:06:42 crc kubenswrapper[4559]: I1123 07:06:42.153449 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46r2s"] Nov 23 07:06:42 crc kubenswrapper[4559]: I1123 07:06:42.357398 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46r2s" event={"ID":"fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29","Type":"ContainerStarted","Data":"ee62e99090e04c68faddd1111da310dd28da264e517148d160f47ce7082f7d79"} Nov 23 07:06:43 crc kubenswrapper[4559]: I1123 07:06:43.366364 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46r2s" event={"ID":"fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29","Type":"ContainerStarted","Data":"33a66b9f98e41849a60a6985b3994a163b33ef5d5094c8d08c10a256efb88786"} Nov 23 07:06:43 crc kubenswrapper[4559]: I1123 07:06:43.380325 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46r2s" podStartSLOduration=1.8475667310000001 podStartE2EDuration="2.380308696s" podCreationTimestamp="2025-11-23 07:06:41 +0000 UTC" firstStartedPulling="2025-11-23 07:06:42.157955853 +0000 UTC m=+1304.179941468" lastFinishedPulling="2025-11-23 07:06:42.690697819 +0000 UTC m=+1304.712683433" observedRunningTime="2025-11-23 07:06:43.378628145 +0000 UTC m=+1305.400613759" watchObservedRunningTime="2025-11-23 07:06:43.380308696 +0000 UTC m=+1305.402294310" Nov 23 07:08:13 crc kubenswrapper[4559]: I1123 07:08:13.192449 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vbkzt"] Nov 23 07:08:13 crc kubenswrapper[4559]: I1123 07:08:13.194558 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vbkzt" Nov 23 07:08:13 crc kubenswrapper[4559]: I1123 07:08:13.204301 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3115257-2c0b-46f2-a9f8-c5f74541c6bf-catalog-content\") pod \"certified-operators-vbkzt\" (UID: \"f3115257-2c0b-46f2-a9f8-c5f74541c6bf\") " pod="openshift-marketplace/certified-operators-vbkzt" Nov 23 07:08:13 crc kubenswrapper[4559]: I1123 07:08:13.204405 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3115257-2c0b-46f2-a9f8-c5f74541c6bf-utilities\") pod \"certified-operators-vbkzt\" (UID: \"f3115257-2c0b-46f2-a9f8-c5f74541c6bf\") " pod="openshift-marketplace/certified-operators-vbkzt" Nov 23 07:08:13 crc kubenswrapper[4559]: I1123 07:08:13.204482 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxbp4\" (UniqueName: \"kubernetes.io/projected/f3115257-2c0b-46f2-a9f8-c5f74541c6bf-kube-api-access-sxbp4\") pod \"certified-operators-vbkzt\" (UID: \"f3115257-2c0b-46f2-a9f8-c5f74541c6bf\") " pod="openshift-marketplace/certified-operators-vbkzt" Nov 23 07:08:13 crc kubenswrapper[4559]: I1123 07:08:13.205436 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vbkzt"] Nov 23 07:08:13 crc kubenswrapper[4559]: I1123 07:08:13.305692 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxbp4\" (UniqueName: \"kubernetes.io/projected/f3115257-2c0b-46f2-a9f8-c5f74541c6bf-kube-api-access-sxbp4\") pod \"certified-operators-vbkzt\" (UID: \"f3115257-2c0b-46f2-a9f8-c5f74541c6bf\") " pod="openshift-marketplace/certified-operators-vbkzt" Nov 23 07:08:13 crc kubenswrapper[4559]: I1123 07:08:13.305855 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3115257-2c0b-46f2-a9f8-c5f74541c6bf-catalog-content\") pod \"certified-operators-vbkzt\" (UID: \"f3115257-2c0b-46f2-a9f8-c5f74541c6bf\") " pod="openshift-marketplace/certified-operators-vbkzt" Nov 23 07:08:13 crc kubenswrapper[4559]: I1123 07:08:13.305909 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3115257-2c0b-46f2-a9f8-c5f74541c6bf-utilities\") pod \"certified-operators-vbkzt\" (UID: \"f3115257-2c0b-46f2-a9f8-c5f74541c6bf\") " pod="openshift-marketplace/certified-operators-vbkzt" Nov 23 07:08:13 crc kubenswrapper[4559]: I1123 07:08:13.306318 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3115257-2c0b-46f2-a9f8-c5f74541c6bf-catalog-content\") pod \"certified-operators-vbkzt\" (UID: \"f3115257-2c0b-46f2-a9f8-c5f74541c6bf\") " pod="openshift-marketplace/certified-operators-vbkzt" Nov 23 07:08:13 crc kubenswrapper[4559]: I1123 07:08:13.306717 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3115257-2c0b-46f2-a9f8-c5f74541c6bf-utilities\") pod \"certified-operators-vbkzt\" (UID: \"f3115257-2c0b-46f2-a9f8-c5f74541c6bf\") " pod="openshift-marketplace/certified-operators-vbkzt" Nov 23 07:08:13 crc kubenswrapper[4559]: I1123 07:08:13.322445 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxbp4\" (UniqueName: \"kubernetes.io/projected/f3115257-2c0b-46f2-a9f8-c5f74541c6bf-kube-api-access-sxbp4\") pod \"certified-operators-vbkzt\" (UID: \"f3115257-2c0b-46f2-a9f8-c5f74541c6bf\") " pod="openshift-marketplace/certified-operators-vbkzt" Nov 23 07:08:13 crc kubenswrapper[4559]: I1123 07:08:13.511758 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vbkzt" Nov 23 07:08:13 crc kubenswrapper[4559]: I1123 07:08:13.925988 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vbkzt"] Nov 23 07:08:13 crc kubenswrapper[4559]: W1123 07:08:13.928518 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf3115257_2c0b_46f2_a9f8_c5f74541c6bf.slice/crio-24bd284851dd030d2c77a0ee059ff7c35cbcb7911b143744661d31f5053b6e9b WatchSource:0}: Error finding container 24bd284851dd030d2c77a0ee059ff7c35cbcb7911b143744661d31f5053b6e9b: Status 404 returned error can't find the container with id 24bd284851dd030d2c77a0ee059ff7c35cbcb7911b143744661d31f5053b6e9b Nov 23 07:08:14 crc kubenswrapper[4559]: I1123 07:08:14.033901 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbkzt" event={"ID":"f3115257-2c0b-46f2-a9f8-c5f74541c6bf","Type":"ContainerStarted","Data":"24bd284851dd030d2c77a0ee059ff7c35cbcb7911b143744661d31f5053b6e9b"} Nov 23 07:08:15 crc kubenswrapper[4559]: I1123 07:08:15.043021 4559 generic.go:334] "Generic (PLEG): container finished" podID="f3115257-2c0b-46f2-a9f8-c5f74541c6bf" containerID="42a577e1ffc89c36d6224513fc0cd44b7a224aeffcff57f36b4649fba5b59ca1" exitCode=0 Nov 23 07:08:15 crc kubenswrapper[4559]: I1123 07:08:15.043072 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbkzt" event={"ID":"f3115257-2c0b-46f2-a9f8-c5f74541c6bf","Type":"ContainerDied","Data":"42a577e1ffc89c36d6224513fc0cd44b7a224aeffcff57f36b4649fba5b59ca1"} Nov 23 07:08:16 crc kubenswrapper[4559]: I1123 07:08:16.051937 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbkzt" event={"ID":"f3115257-2c0b-46f2-a9f8-c5f74541c6bf","Type":"ContainerStarted","Data":"675ba79d3914c4a5fd47d0e86fe79f6226a09d54c30b66700c3b8da720cb14a9"} Nov 23 07:08:17 crc kubenswrapper[4559]: I1123 07:08:17.069211 4559 generic.go:334] "Generic (PLEG): container finished" podID="f3115257-2c0b-46f2-a9f8-c5f74541c6bf" containerID="675ba79d3914c4a5fd47d0e86fe79f6226a09d54c30b66700c3b8da720cb14a9" exitCode=0 Nov 23 07:08:17 crc kubenswrapper[4559]: I1123 07:08:17.069987 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbkzt" event={"ID":"f3115257-2c0b-46f2-a9f8-c5f74541c6bf","Type":"ContainerDied","Data":"675ba79d3914c4a5fd47d0e86fe79f6226a09d54c30b66700c3b8da720cb14a9"} Nov 23 07:08:18 crc kubenswrapper[4559]: I1123 07:08:18.077447 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbkzt" event={"ID":"f3115257-2c0b-46f2-a9f8-c5f74541c6bf","Type":"ContainerStarted","Data":"2f158cfe006329e37aed146137bd16513305bc64eb8059754852f20da0de8f2e"} Nov 23 07:08:18 crc kubenswrapper[4559]: I1123 07:08:18.096296 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vbkzt" podStartSLOduration=2.520119286 podStartE2EDuration="5.096282232s" podCreationTimestamp="2025-11-23 07:08:13 +0000 UTC" firstStartedPulling="2025-11-23 07:08:15.045218084 +0000 UTC m=+1397.067203697" lastFinishedPulling="2025-11-23 07:08:17.62138103 +0000 UTC m=+1399.643366643" observedRunningTime="2025-11-23 07:08:18.090549823 +0000 UTC m=+1400.112535438" watchObservedRunningTime="2025-11-23 07:08:18.096282232 +0000 UTC m=+1400.118267846" Nov 23 07:08:20 crc kubenswrapper[4559]: I1123 07:08:20.092554 4559 generic.go:334] "Generic (PLEG): container finished" podID="fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29" containerID="33a66b9f98e41849a60a6985b3994a163b33ef5d5094c8d08c10a256efb88786" exitCode=0 Nov 23 07:08:20 crc kubenswrapper[4559]: I1123 07:08:20.092668 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46r2s" event={"ID":"fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29","Type":"ContainerDied","Data":"33a66b9f98e41849a60a6985b3994a163b33ef5d5094c8d08c10a256efb88786"} Nov 23 07:08:21 crc kubenswrapper[4559]: I1123 07:08:21.386737 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46r2s" Nov 23 07:08:21 crc kubenswrapper[4559]: I1123 07:08:21.526823 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-st6jn\" (UniqueName: \"kubernetes.io/projected/fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29-kube-api-access-st6jn\") pod \"fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29\" (UID: \"fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29\") " Nov 23 07:08:21 crc kubenswrapper[4559]: I1123 07:08:21.526967 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29-ssh-key\") pod \"fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29\" (UID: \"fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29\") " Nov 23 07:08:21 crc kubenswrapper[4559]: I1123 07:08:21.526990 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29-inventory\") pod \"fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29\" (UID: \"fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29\") " Nov 23 07:08:21 crc kubenswrapper[4559]: I1123 07:08:21.534766 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29-kube-api-access-st6jn" (OuterVolumeSpecName: "kube-api-access-st6jn") pod "fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29" (UID: "fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29"). InnerVolumeSpecName "kube-api-access-st6jn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:08:21 crc kubenswrapper[4559]: I1123 07:08:21.547097 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29-inventory" (OuterVolumeSpecName: "inventory") pod "fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29" (UID: "fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:08:21 crc kubenswrapper[4559]: I1123 07:08:21.547375 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29" (UID: "fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:08:21 crc kubenswrapper[4559]: I1123 07:08:21.629206 4559 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:21 crc kubenswrapper[4559]: I1123 07:08:21.629232 4559 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:21 crc kubenswrapper[4559]: I1123 07:08:21.629241 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-st6jn\" (UniqueName: \"kubernetes.io/projected/fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29-kube-api-access-st6jn\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:22 crc kubenswrapper[4559]: I1123 07:08:22.107980 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46r2s" event={"ID":"fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29","Type":"ContainerDied","Data":"ee62e99090e04c68faddd1111da310dd28da264e517148d160f47ce7082f7d79"} Nov 23 07:08:22 crc kubenswrapper[4559]: I1123 07:08:22.108017 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee62e99090e04c68faddd1111da310dd28da264e517148d160f47ce7082f7d79" Nov 23 07:08:22 crc kubenswrapper[4559]: I1123 07:08:22.108023 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46r2s" Nov 23 07:08:22 crc kubenswrapper[4559]: I1123 07:08:22.159262 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5"] Nov 23 07:08:22 crc kubenswrapper[4559]: E1123 07:08:22.159660 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 23 07:08:22 crc kubenswrapper[4559]: I1123 07:08:22.159678 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 23 07:08:22 crc kubenswrapper[4559]: I1123 07:08:22.159847 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 23 07:08:22 crc kubenswrapper[4559]: I1123 07:08:22.160426 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5" Nov 23 07:08:22 crc kubenswrapper[4559]: I1123 07:08:22.162263 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 07:08:22 crc kubenswrapper[4559]: I1123 07:08:22.162422 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 07:08:22 crc kubenswrapper[4559]: I1123 07:08:22.162464 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 07:08:22 crc kubenswrapper[4559]: I1123 07:08:22.163719 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mj6c" Nov 23 07:08:22 crc kubenswrapper[4559]: I1123 07:08:22.165606 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5"] Nov 23 07:08:22 crc kubenswrapper[4559]: I1123 07:08:22.338007 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/45c516df-85dd-4a43-94ef-ea1381622c06-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5\" (UID: \"45c516df-85dd-4a43-94ef-ea1381622c06\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5" Nov 23 07:08:22 crc kubenswrapper[4559]: I1123 07:08:22.338225 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/45c516df-85dd-4a43-94ef-ea1381622c06-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5\" (UID: \"45c516df-85dd-4a43-94ef-ea1381622c06\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5" Nov 23 07:08:22 crc kubenswrapper[4559]: I1123 07:08:22.338254 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8ncb\" (UniqueName: \"kubernetes.io/projected/45c516df-85dd-4a43-94ef-ea1381622c06-kube-api-access-m8ncb\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5\" (UID: \"45c516df-85dd-4a43-94ef-ea1381622c06\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5" Nov 23 07:08:22 crc kubenswrapper[4559]: I1123 07:08:22.439508 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/45c516df-85dd-4a43-94ef-ea1381622c06-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5\" (UID: \"45c516df-85dd-4a43-94ef-ea1381622c06\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5" Nov 23 07:08:22 crc kubenswrapper[4559]: I1123 07:08:22.439567 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/45c516df-85dd-4a43-94ef-ea1381622c06-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5\" (UID: \"45c516df-85dd-4a43-94ef-ea1381622c06\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5" Nov 23 07:08:22 crc kubenswrapper[4559]: I1123 07:08:22.439603 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8ncb\" (UniqueName: \"kubernetes.io/projected/45c516df-85dd-4a43-94ef-ea1381622c06-kube-api-access-m8ncb\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5\" (UID: \"45c516df-85dd-4a43-94ef-ea1381622c06\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5" Nov 23 07:08:22 crc kubenswrapper[4559]: I1123 07:08:22.442465 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/45c516df-85dd-4a43-94ef-ea1381622c06-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5\" (UID: \"45c516df-85dd-4a43-94ef-ea1381622c06\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5" Nov 23 07:08:22 crc kubenswrapper[4559]: I1123 07:08:22.442517 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/45c516df-85dd-4a43-94ef-ea1381622c06-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5\" (UID: \"45c516df-85dd-4a43-94ef-ea1381622c06\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5" Nov 23 07:08:22 crc kubenswrapper[4559]: I1123 07:08:22.452765 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8ncb\" (UniqueName: \"kubernetes.io/projected/45c516df-85dd-4a43-94ef-ea1381622c06-kube-api-access-m8ncb\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5\" (UID: \"45c516df-85dd-4a43-94ef-ea1381622c06\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5" Nov 23 07:08:22 crc kubenswrapper[4559]: I1123 07:08:22.472638 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5" Nov 23 07:08:22 crc kubenswrapper[4559]: I1123 07:08:22.884776 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5"] Nov 23 07:08:23 crc kubenswrapper[4559]: I1123 07:08:23.115274 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5" event={"ID":"45c516df-85dd-4a43-94ef-ea1381622c06","Type":"ContainerStarted","Data":"d490683b13a70ef53e47b4a0208933c3b588f4a064aeb5968e6a68a834bae20a"} Nov 23 07:08:23 crc kubenswrapper[4559]: I1123 07:08:23.512536 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vbkzt" Nov 23 07:08:23 crc kubenswrapper[4559]: I1123 07:08:23.512715 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vbkzt" Nov 23 07:08:23 crc kubenswrapper[4559]: I1123 07:08:23.545178 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vbkzt" Nov 23 07:08:24 crc kubenswrapper[4559]: I1123 07:08:24.122693 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5" event={"ID":"45c516df-85dd-4a43-94ef-ea1381622c06","Type":"ContainerStarted","Data":"e49821d0092fdae1e63db0c0a13a10e0a8843316db83f6e7b812cc247a31aa44"} Nov 23 07:08:24 crc kubenswrapper[4559]: I1123 07:08:24.156192 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5" podStartSLOduration=1.600809186 podStartE2EDuration="2.156176892s" podCreationTimestamp="2025-11-23 07:08:22 +0000 UTC" firstStartedPulling="2025-11-23 07:08:22.887003961 +0000 UTC m=+1404.908989575" lastFinishedPulling="2025-11-23 07:08:23.442371668 +0000 UTC m=+1405.464357281" observedRunningTime="2025-11-23 07:08:24.148748126 +0000 UTC m=+1406.170733740" watchObservedRunningTime="2025-11-23 07:08:24.156176892 +0000 UTC m=+1406.178162506" Nov 23 07:08:24 crc kubenswrapper[4559]: I1123 07:08:24.159924 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vbkzt" Nov 23 07:08:24 crc kubenswrapper[4559]: I1123 07:08:24.191550 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vbkzt"] Nov 23 07:08:26 crc kubenswrapper[4559]: I1123 07:08:26.134963 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vbkzt" podUID="f3115257-2c0b-46f2-a9f8-c5f74541c6bf" containerName="registry-server" containerID="cri-o://2f158cfe006329e37aed146137bd16513305bc64eb8059754852f20da0de8f2e" gracePeriod=2 Nov 23 07:08:26 crc kubenswrapper[4559]: I1123 07:08:26.167411 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:08:26 crc kubenswrapper[4559]: I1123 07:08:26.167454 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:08:26 crc kubenswrapper[4559]: I1123 07:08:26.474507 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vbkzt" Nov 23 07:08:26 crc kubenswrapper[4559]: I1123 07:08:26.599564 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3115257-2c0b-46f2-a9f8-c5f74541c6bf-utilities\") pod \"f3115257-2c0b-46f2-a9f8-c5f74541c6bf\" (UID: \"f3115257-2c0b-46f2-a9f8-c5f74541c6bf\") " Nov 23 07:08:26 crc kubenswrapper[4559]: I1123 07:08:26.599610 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3115257-2c0b-46f2-a9f8-c5f74541c6bf-catalog-content\") pod \"f3115257-2c0b-46f2-a9f8-c5f74541c6bf\" (UID: \"f3115257-2c0b-46f2-a9f8-c5f74541c6bf\") " Nov 23 07:08:26 crc kubenswrapper[4559]: I1123 07:08:26.599661 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxbp4\" (UniqueName: \"kubernetes.io/projected/f3115257-2c0b-46f2-a9f8-c5f74541c6bf-kube-api-access-sxbp4\") pod \"f3115257-2c0b-46f2-a9f8-c5f74541c6bf\" (UID: \"f3115257-2c0b-46f2-a9f8-c5f74541c6bf\") " Nov 23 07:08:26 crc kubenswrapper[4559]: I1123 07:08:26.600509 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3115257-2c0b-46f2-a9f8-c5f74541c6bf-utilities" (OuterVolumeSpecName: "utilities") pod "f3115257-2c0b-46f2-a9f8-c5f74541c6bf" (UID: "f3115257-2c0b-46f2-a9f8-c5f74541c6bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:08:26 crc kubenswrapper[4559]: I1123 07:08:26.604462 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3115257-2c0b-46f2-a9f8-c5f74541c6bf-kube-api-access-sxbp4" (OuterVolumeSpecName: "kube-api-access-sxbp4") pod "f3115257-2c0b-46f2-a9f8-c5f74541c6bf" (UID: "f3115257-2c0b-46f2-a9f8-c5f74541c6bf"). InnerVolumeSpecName "kube-api-access-sxbp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:08:26 crc kubenswrapper[4559]: I1123 07:08:26.701410 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3115257-2c0b-46f2-a9f8-c5f74541c6bf-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:26 crc kubenswrapper[4559]: I1123 07:08:26.701441 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxbp4\" (UniqueName: \"kubernetes.io/projected/f3115257-2c0b-46f2-a9f8-c5f74541c6bf-kube-api-access-sxbp4\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:26 crc kubenswrapper[4559]: I1123 07:08:26.715570 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3115257-2c0b-46f2-a9f8-c5f74541c6bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f3115257-2c0b-46f2-a9f8-c5f74541c6bf" (UID: "f3115257-2c0b-46f2-a9f8-c5f74541c6bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:08:26 crc kubenswrapper[4559]: I1123 07:08:26.802696 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3115257-2c0b-46f2-a9f8-c5f74541c6bf-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:08:27 crc kubenswrapper[4559]: I1123 07:08:27.143827 4559 generic.go:334] "Generic (PLEG): container finished" podID="f3115257-2c0b-46f2-a9f8-c5f74541c6bf" containerID="2f158cfe006329e37aed146137bd16513305bc64eb8059754852f20da0de8f2e" exitCode=0 Nov 23 07:08:27 crc kubenswrapper[4559]: I1123 07:08:27.143862 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbkzt" event={"ID":"f3115257-2c0b-46f2-a9f8-c5f74541c6bf","Type":"ContainerDied","Data":"2f158cfe006329e37aed146137bd16513305bc64eb8059754852f20da0de8f2e"} Nov 23 07:08:27 crc kubenswrapper[4559]: I1123 07:08:27.143888 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vbkzt" Nov 23 07:08:27 crc kubenswrapper[4559]: I1123 07:08:27.143888 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbkzt" event={"ID":"f3115257-2c0b-46f2-a9f8-c5f74541c6bf","Type":"ContainerDied","Data":"24bd284851dd030d2c77a0ee059ff7c35cbcb7911b143744661d31f5053b6e9b"} Nov 23 07:08:27 crc kubenswrapper[4559]: I1123 07:08:27.143922 4559 scope.go:117] "RemoveContainer" containerID="2f158cfe006329e37aed146137bd16513305bc64eb8059754852f20da0de8f2e" Nov 23 07:08:27 crc kubenswrapper[4559]: I1123 07:08:27.168975 4559 scope.go:117] "RemoveContainer" containerID="675ba79d3914c4a5fd47d0e86fe79f6226a09d54c30b66700c3b8da720cb14a9" Nov 23 07:08:27 crc kubenswrapper[4559]: I1123 07:08:27.171544 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vbkzt"] Nov 23 07:08:27 crc kubenswrapper[4559]: I1123 07:08:27.177512 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vbkzt"] Nov 23 07:08:27 crc kubenswrapper[4559]: I1123 07:08:27.202985 4559 scope.go:117] "RemoveContainer" containerID="42a577e1ffc89c36d6224513fc0cd44b7a224aeffcff57f36b4649fba5b59ca1" Nov 23 07:08:27 crc kubenswrapper[4559]: I1123 07:08:27.218675 4559 scope.go:117] "RemoveContainer" containerID="2f158cfe006329e37aed146137bd16513305bc64eb8059754852f20da0de8f2e" Nov 23 07:08:27 crc kubenswrapper[4559]: E1123 07:08:27.219068 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f158cfe006329e37aed146137bd16513305bc64eb8059754852f20da0de8f2e\": container with ID starting with 2f158cfe006329e37aed146137bd16513305bc64eb8059754852f20da0de8f2e not found: ID does not exist" containerID="2f158cfe006329e37aed146137bd16513305bc64eb8059754852f20da0de8f2e" Nov 23 07:08:27 crc kubenswrapper[4559]: I1123 07:08:27.219105 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f158cfe006329e37aed146137bd16513305bc64eb8059754852f20da0de8f2e"} err="failed to get container status \"2f158cfe006329e37aed146137bd16513305bc64eb8059754852f20da0de8f2e\": rpc error: code = NotFound desc = could not find container \"2f158cfe006329e37aed146137bd16513305bc64eb8059754852f20da0de8f2e\": container with ID starting with 2f158cfe006329e37aed146137bd16513305bc64eb8059754852f20da0de8f2e not found: ID does not exist" Nov 23 07:08:27 crc kubenswrapper[4559]: I1123 07:08:27.219127 4559 scope.go:117] "RemoveContainer" containerID="675ba79d3914c4a5fd47d0e86fe79f6226a09d54c30b66700c3b8da720cb14a9" Nov 23 07:08:27 crc kubenswrapper[4559]: E1123 07:08:27.219479 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"675ba79d3914c4a5fd47d0e86fe79f6226a09d54c30b66700c3b8da720cb14a9\": container with ID starting with 675ba79d3914c4a5fd47d0e86fe79f6226a09d54c30b66700c3b8da720cb14a9 not found: ID does not exist" containerID="675ba79d3914c4a5fd47d0e86fe79f6226a09d54c30b66700c3b8da720cb14a9" Nov 23 07:08:27 crc kubenswrapper[4559]: I1123 07:08:27.219515 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"675ba79d3914c4a5fd47d0e86fe79f6226a09d54c30b66700c3b8da720cb14a9"} err="failed to get container status \"675ba79d3914c4a5fd47d0e86fe79f6226a09d54c30b66700c3b8da720cb14a9\": rpc error: code = NotFound desc = could not find container \"675ba79d3914c4a5fd47d0e86fe79f6226a09d54c30b66700c3b8da720cb14a9\": container with ID starting with 675ba79d3914c4a5fd47d0e86fe79f6226a09d54c30b66700c3b8da720cb14a9 not found: ID does not exist" Nov 23 07:08:27 crc kubenswrapper[4559]: I1123 07:08:27.219539 4559 scope.go:117] "RemoveContainer" containerID="42a577e1ffc89c36d6224513fc0cd44b7a224aeffcff57f36b4649fba5b59ca1" Nov 23 07:08:27 crc kubenswrapper[4559]: E1123 07:08:27.219891 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42a577e1ffc89c36d6224513fc0cd44b7a224aeffcff57f36b4649fba5b59ca1\": container with ID starting with 42a577e1ffc89c36d6224513fc0cd44b7a224aeffcff57f36b4649fba5b59ca1 not found: ID does not exist" containerID="42a577e1ffc89c36d6224513fc0cd44b7a224aeffcff57f36b4649fba5b59ca1" Nov 23 07:08:27 crc kubenswrapper[4559]: I1123 07:08:27.219939 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42a577e1ffc89c36d6224513fc0cd44b7a224aeffcff57f36b4649fba5b59ca1"} err="failed to get container status \"42a577e1ffc89c36d6224513fc0cd44b7a224aeffcff57f36b4649fba5b59ca1\": rpc error: code = NotFound desc = could not find container \"42a577e1ffc89c36d6224513fc0cd44b7a224aeffcff57f36b4649fba5b59ca1\": container with ID starting with 42a577e1ffc89c36d6224513fc0cd44b7a224aeffcff57f36b4649fba5b59ca1 not found: ID does not exist" Nov 23 07:08:28 crc kubenswrapper[4559]: I1123 07:08:28.281488 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3115257-2c0b-46f2-a9f8-c5f74541c6bf" path="/var/lib/kubelet/pods/f3115257-2c0b-46f2-a9f8-c5f74541c6bf/volumes" Nov 23 07:08:31 crc kubenswrapper[4559]: I1123 07:08:31.027892 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-556xz"] Nov 23 07:08:31 crc kubenswrapper[4559]: I1123 07:08:31.034902 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-48ksj"] Nov 23 07:08:31 crc kubenswrapper[4559]: I1123 07:08:31.041435 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-c1ff-account-create-ngpwh"] Nov 23 07:08:31 crc kubenswrapper[4559]: I1123 07:08:31.048668 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-9b47-account-create-br9bd"] Nov 23 07:08:31 crc kubenswrapper[4559]: I1123 07:08:31.053658 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-556xz"] Nov 23 07:08:31 crc kubenswrapper[4559]: I1123 07:08:31.058526 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-48ksj"] Nov 23 07:08:31 crc kubenswrapper[4559]: I1123 07:08:31.063133 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-c1ff-account-create-ngpwh"] Nov 23 07:08:31 crc kubenswrapper[4559]: I1123 07:08:31.068009 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-9b47-account-create-br9bd"] Nov 23 07:08:32 crc kubenswrapper[4559]: I1123 07:08:32.281294 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2efe0b7a-32c9-40b6-a933-014c9561b25e" path="/var/lib/kubelet/pods/2efe0b7a-32c9-40b6-a933-014c9561b25e/volumes" Nov 23 07:08:32 crc kubenswrapper[4559]: I1123 07:08:32.281829 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a81225a3-0928-42f9-bdc8-22375a649775" path="/var/lib/kubelet/pods/a81225a3-0928-42f9-bdc8-22375a649775/volumes" Nov 23 07:08:32 crc kubenswrapper[4559]: I1123 07:08:32.282340 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de1904f5-a9a6-4bc2-b106-71cbc351650b" path="/var/lib/kubelet/pods/de1904f5-a9a6-4bc2-b106-71cbc351650b/volumes" Nov 23 07:08:32 crc kubenswrapper[4559]: I1123 07:08:32.282820 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f26678a1-0d1d-40c6-9eef-15f8ff08caab" path="/var/lib/kubelet/pods/f26678a1-0d1d-40c6-9eef-15f8ff08caab/volumes" Nov 23 07:08:36 crc kubenswrapper[4559]: I1123 07:08:36.019428 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-njl4x"] Nov 23 07:08:36 crc kubenswrapper[4559]: I1123 07:08:36.025744 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-7b13-account-create-mzr4r"] Nov 23 07:08:36 crc kubenswrapper[4559]: I1123 07:08:36.032105 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-njl4x"] Nov 23 07:08:36 crc kubenswrapper[4559]: I1123 07:08:36.037143 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-7b13-account-create-mzr4r"] Nov 23 07:08:36 crc kubenswrapper[4559]: I1123 07:08:36.281764 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37737b2f-1fb8-4028-a7a2-134fb6e0b991" path="/var/lib/kubelet/pods/37737b2f-1fb8-4028-a7a2-134fb6e0b991/volumes" Nov 23 07:08:36 crc kubenswrapper[4559]: I1123 07:08:36.282299 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5018f2c-cf19-4991-a3ad-5909128d9de9" path="/var/lib/kubelet/pods/f5018f2c-cf19-4991-a3ad-5909128d9de9/volumes" Nov 23 07:08:56 crc kubenswrapper[4559]: I1123 07:08:56.166859 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:08:56 crc kubenswrapper[4559]: I1123 07:08:56.167861 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:09:02 crc kubenswrapper[4559]: I1123 07:09:02.024459 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-dwbhb"] Nov 23 07:09:02 crc kubenswrapper[4559]: I1123 07:09:02.029662 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-dwbhb"] Nov 23 07:09:02 crc kubenswrapper[4559]: I1123 07:09:02.280935 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5cc366d-e466-476d-a318-1582a4de4a93" path="/var/lib/kubelet/pods/a5cc366d-e466-476d-a318-1582a4de4a93/volumes" Nov 23 07:09:09 crc kubenswrapper[4559]: I1123 07:09:09.827545 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hhqzb"] Nov 23 07:09:09 crc kubenswrapper[4559]: E1123 07:09:09.829788 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3115257-2c0b-46f2-a9f8-c5f74541c6bf" containerName="registry-server" Nov 23 07:09:09 crc kubenswrapper[4559]: I1123 07:09:09.829864 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3115257-2c0b-46f2-a9f8-c5f74541c6bf" containerName="registry-server" Nov 23 07:09:09 crc kubenswrapper[4559]: E1123 07:09:09.829919 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3115257-2c0b-46f2-a9f8-c5f74541c6bf" containerName="extract-utilities" Nov 23 07:09:09 crc kubenswrapper[4559]: I1123 07:09:09.829968 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3115257-2c0b-46f2-a9f8-c5f74541c6bf" containerName="extract-utilities" Nov 23 07:09:09 crc kubenswrapper[4559]: E1123 07:09:09.830050 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3115257-2c0b-46f2-a9f8-c5f74541c6bf" containerName="extract-content" Nov 23 07:09:09 crc kubenswrapper[4559]: I1123 07:09:09.830096 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3115257-2c0b-46f2-a9f8-c5f74541c6bf" containerName="extract-content" Nov 23 07:09:09 crc kubenswrapper[4559]: I1123 07:09:09.830288 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3115257-2c0b-46f2-a9f8-c5f74541c6bf" containerName="registry-server" Nov 23 07:09:09 crc kubenswrapper[4559]: I1123 07:09:09.831454 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hhqzb" Nov 23 07:09:09 crc kubenswrapper[4559]: I1123 07:09:09.838106 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hhqzb"] Nov 23 07:09:09 crc kubenswrapper[4559]: I1123 07:09:09.931752 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da93db43-37fc-40c9-a2e2-9916e26ef920-catalog-content\") pod \"community-operators-hhqzb\" (UID: \"da93db43-37fc-40c9-a2e2-9916e26ef920\") " pod="openshift-marketplace/community-operators-hhqzb" Nov 23 07:09:09 crc kubenswrapper[4559]: I1123 07:09:09.932035 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s5dc\" (UniqueName: \"kubernetes.io/projected/da93db43-37fc-40c9-a2e2-9916e26ef920-kube-api-access-2s5dc\") pod \"community-operators-hhqzb\" (UID: \"da93db43-37fc-40c9-a2e2-9916e26ef920\") " pod="openshift-marketplace/community-operators-hhqzb" Nov 23 07:09:09 crc kubenswrapper[4559]: I1123 07:09:09.932165 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da93db43-37fc-40c9-a2e2-9916e26ef920-utilities\") pod \"community-operators-hhqzb\" (UID: \"da93db43-37fc-40c9-a2e2-9916e26ef920\") " pod="openshift-marketplace/community-operators-hhqzb" Nov 23 07:09:10 crc kubenswrapper[4559]: I1123 07:09:10.034111 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da93db43-37fc-40c9-a2e2-9916e26ef920-catalog-content\") pod \"community-operators-hhqzb\" (UID: \"da93db43-37fc-40c9-a2e2-9916e26ef920\") " pod="openshift-marketplace/community-operators-hhqzb" Nov 23 07:09:10 crc kubenswrapper[4559]: I1123 07:09:10.034218 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s5dc\" (UniqueName: \"kubernetes.io/projected/da93db43-37fc-40c9-a2e2-9916e26ef920-kube-api-access-2s5dc\") pod \"community-operators-hhqzb\" (UID: \"da93db43-37fc-40c9-a2e2-9916e26ef920\") " pod="openshift-marketplace/community-operators-hhqzb" Nov 23 07:09:10 crc kubenswrapper[4559]: I1123 07:09:10.034257 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da93db43-37fc-40c9-a2e2-9916e26ef920-utilities\") pod \"community-operators-hhqzb\" (UID: \"da93db43-37fc-40c9-a2e2-9916e26ef920\") " pod="openshift-marketplace/community-operators-hhqzb" Nov 23 07:09:10 crc kubenswrapper[4559]: I1123 07:09:10.034654 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da93db43-37fc-40c9-a2e2-9916e26ef920-catalog-content\") pod \"community-operators-hhqzb\" (UID: \"da93db43-37fc-40c9-a2e2-9916e26ef920\") " pod="openshift-marketplace/community-operators-hhqzb" Nov 23 07:09:10 crc kubenswrapper[4559]: I1123 07:09:10.034709 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da93db43-37fc-40c9-a2e2-9916e26ef920-utilities\") pod \"community-operators-hhqzb\" (UID: \"da93db43-37fc-40c9-a2e2-9916e26ef920\") " pod="openshift-marketplace/community-operators-hhqzb" Nov 23 07:09:10 crc kubenswrapper[4559]: I1123 07:09:10.050517 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s5dc\" (UniqueName: \"kubernetes.io/projected/da93db43-37fc-40c9-a2e2-9916e26ef920-kube-api-access-2s5dc\") pod \"community-operators-hhqzb\" (UID: \"da93db43-37fc-40c9-a2e2-9916e26ef920\") " pod="openshift-marketplace/community-operators-hhqzb" Nov 23 07:09:10 crc kubenswrapper[4559]: I1123 07:09:10.159698 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hhqzb" Nov 23 07:09:10 crc kubenswrapper[4559]: I1123 07:09:10.585100 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hhqzb"] Nov 23 07:09:11 crc kubenswrapper[4559]: I1123 07:09:11.436108 4559 generic.go:334] "Generic (PLEG): container finished" podID="da93db43-37fc-40c9-a2e2-9916e26ef920" containerID="b2766e8e6f8c3caad942345a56a48163ac88d207a21f42e35c1ab1902004e1b0" exitCode=0 Nov 23 07:09:11 crc kubenswrapper[4559]: I1123 07:09:11.436181 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hhqzb" event={"ID":"da93db43-37fc-40c9-a2e2-9916e26ef920","Type":"ContainerDied","Data":"b2766e8e6f8c3caad942345a56a48163ac88d207a21f42e35c1ab1902004e1b0"} Nov 23 07:09:11 crc kubenswrapper[4559]: I1123 07:09:11.436428 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hhqzb" event={"ID":"da93db43-37fc-40c9-a2e2-9916e26ef920","Type":"ContainerStarted","Data":"ea59ceea00fb775adc30d702bfed181d19eaa2bed2339b745075c6c0b874f260"} Nov 23 07:09:11 crc kubenswrapper[4559]: I1123 07:09:11.438325 4559 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 07:09:12 crc kubenswrapper[4559]: I1123 07:09:12.218709 4559 scope.go:117] "RemoveContainer" containerID="dc01e112a1fab31fb727f9b68a2057b14624fbf2f00723615b543be9cc67de1e" Nov 23 07:09:12 crc kubenswrapper[4559]: I1123 07:09:12.237242 4559 scope.go:117] "RemoveContainer" containerID="447e3db48c4b79fa32cad92f618795df76a33cfc391fb7dadf972e3631cb3360" Nov 23 07:09:12 crc kubenswrapper[4559]: I1123 07:09:12.267377 4559 scope.go:117] "RemoveContainer" containerID="28c045ec04f1f1fbec6423600a427ac6ae17f3d9f67ebec610f72f1a248b0593" Nov 23 07:09:12 crc kubenswrapper[4559]: I1123 07:09:12.312734 4559 scope.go:117] "RemoveContainer" containerID="f3cbd5748e3000217e906faa8173d52b20792663728a89a32049aba6b07e3e6b" Nov 23 07:09:12 crc kubenswrapper[4559]: I1123 07:09:12.344300 4559 scope.go:117] "RemoveContainer" containerID="670f508ab528337d97e501177b1ee9f4694a39f89314fc6752425ecda77b578a" Nov 23 07:09:12 crc kubenswrapper[4559]: I1123 07:09:12.433871 4559 scope.go:117] "RemoveContainer" containerID="50e879df85b3e5d509462b0d8abd7c9217e25644c9e956a57311ed210f16d4c2" Nov 23 07:09:12 crc kubenswrapper[4559]: I1123 07:09:12.454431 4559 generic.go:334] "Generic (PLEG): container finished" podID="da93db43-37fc-40c9-a2e2-9916e26ef920" containerID="2e4572001f1666e734759d0badef211fe79d01c9e82dc39425c538b038a9474e" exitCode=0 Nov 23 07:09:12 crc kubenswrapper[4559]: I1123 07:09:12.454502 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hhqzb" event={"ID":"da93db43-37fc-40c9-a2e2-9916e26ef920","Type":"ContainerDied","Data":"2e4572001f1666e734759d0badef211fe79d01c9e82dc39425c538b038a9474e"} Nov 23 07:09:12 crc kubenswrapper[4559]: I1123 07:09:12.455145 4559 scope.go:117] "RemoveContainer" containerID="48f592e4030e6b3434d490cb8ff49b0778303be19fcee13d4246eecff5a75668" Nov 23 07:09:13 crc kubenswrapper[4559]: I1123 07:09:13.463933 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hhqzb" event={"ID":"da93db43-37fc-40c9-a2e2-9916e26ef920","Type":"ContainerStarted","Data":"ad819b987952cdf194b260e62c60987225a8a7d2e40084f66250e41664b1806d"} Nov 23 07:09:13 crc kubenswrapper[4559]: I1123 07:09:13.480438 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hhqzb" podStartSLOduration=2.867443165 podStartE2EDuration="4.480423377s" podCreationTimestamp="2025-11-23 07:09:09 +0000 UTC" firstStartedPulling="2025-11-23 07:09:11.438084617 +0000 UTC m=+1453.460070231" lastFinishedPulling="2025-11-23 07:09:13.05106483 +0000 UTC m=+1455.073050443" observedRunningTime="2025-11-23 07:09:13.477246649 +0000 UTC m=+1455.499232263" watchObservedRunningTime="2025-11-23 07:09:13.480423377 +0000 UTC m=+1455.502408991" Nov 23 07:09:18 crc kubenswrapper[4559]: I1123 07:09:18.020246 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-j4kkb"] Nov 23 07:09:18 crc kubenswrapper[4559]: I1123 07:09:18.027451 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-f4d5-account-create-q6dwz"] Nov 23 07:09:18 crc kubenswrapper[4559]: I1123 07:09:18.033981 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-j4kkb"] Nov 23 07:09:18 crc kubenswrapper[4559]: I1123 07:09:18.040465 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-f4d5-account-create-q6dwz"] Nov 23 07:09:18 crc kubenswrapper[4559]: I1123 07:09:18.283796 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11264a11-4f18-4550-a335-c0fd08428786" path="/var/lib/kubelet/pods/11264a11-4f18-4550-a335-c0fd08428786/volumes" Nov 23 07:09:18 crc kubenswrapper[4559]: I1123 07:09:18.284746 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39400620-4347-45f5-85ee-4bfed29cb5cd" path="/var/lib/kubelet/pods/39400620-4347-45f5-85ee-4bfed29cb5cd/volumes" Nov 23 07:09:19 crc kubenswrapper[4559]: I1123 07:09:19.023452 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-a32f-account-create-h2s4w"] Nov 23 07:09:19 crc kubenswrapper[4559]: I1123 07:09:19.028584 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-6ad9-account-create-bfsj7"] Nov 23 07:09:19 crc kubenswrapper[4559]: I1123 07:09:19.035264 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-2xf6v"] Nov 23 07:09:19 crc kubenswrapper[4559]: I1123 07:09:19.039787 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-65wms"] Nov 23 07:09:19 crc kubenswrapper[4559]: I1123 07:09:19.053140 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-a32f-account-create-h2s4w"] Nov 23 07:09:19 crc kubenswrapper[4559]: I1123 07:09:19.058547 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-6ad9-account-create-bfsj7"] Nov 23 07:09:19 crc kubenswrapper[4559]: I1123 07:09:19.069142 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-65wms"] Nov 23 07:09:19 crc kubenswrapper[4559]: I1123 07:09:19.075733 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-2xf6v"] Nov 23 07:09:20 crc kubenswrapper[4559]: I1123 07:09:20.161077 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hhqzb" Nov 23 07:09:20 crc kubenswrapper[4559]: I1123 07:09:20.161439 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hhqzb" Nov 23 07:09:20 crc kubenswrapper[4559]: I1123 07:09:20.201041 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hhqzb" Nov 23 07:09:20 crc kubenswrapper[4559]: I1123 07:09:20.281526 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="229ac05d-35f6-4775-8298-c61aa1378406" path="/var/lib/kubelet/pods/229ac05d-35f6-4775-8298-c61aa1378406/volumes" Nov 23 07:09:20 crc kubenswrapper[4559]: I1123 07:09:20.282267 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2b9eb8d-f298-4085-9b0e-4c9c929e23c2" path="/var/lib/kubelet/pods/a2b9eb8d-f298-4085-9b0e-4c9c929e23c2/volumes" Nov 23 07:09:20 crc kubenswrapper[4559]: I1123 07:09:20.282903 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6600577-55b0-439d-85c2-dab41fac1775" path="/var/lib/kubelet/pods/c6600577-55b0-439d-85c2-dab41fac1775/volumes" Nov 23 07:09:20 crc kubenswrapper[4559]: I1123 07:09:20.283420 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d08d731c-9663-4eb7-ac6d-8f500dc4b294" path="/var/lib/kubelet/pods/d08d731c-9663-4eb7-ac6d-8f500dc4b294/volumes" Nov 23 07:09:20 crc kubenswrapper[4559]: I1123 07:09:20.545033 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hhqzb" Nov 23 07:09:20 crc kubenswrapper[4559]: I1123 07:09:20.581560 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hhqzb"] Nov 23 07:09:22 crc kubenswrapper[4559]: I1123 07:09:22.526059 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hhqzb" podUID="da93db43-37fc-40c9-a2e2-9916e26ef920" containerName="registry-server" containerID="cri-o://ad819b987952cdf194b260e62c60987225a8a7d2e40084f66250e41664b1806d" gracePeriod=2 Nov 23 07:09:22 crc kubenswrapper[4559]: I1123 07:09:22.872306 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hhqzb" Nov 23 07:09:22 crc kubenswrapper[4559]: I1123 07:09:22.960819 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2s5dc\" (UniqueName: \"kubernetes.io/projected/da93db43-37fc-40c9-a2e2-9916e26ef920-kube-api-access-2s5dc\") pod \"da93db43-37fc-40c9-a2e2-9916e26ef920\" (UID: \"da93db43-37fc-40c9-a2e2-9916e26ef920\") " Nov 23 07:09:22 crc kubenswrapper[4559]: I1123 07:09:22.960899 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da93db43-37fc-40c9-a2e2-9916e26ef920-catalog-content\") pod \"da93db43-37fc-40c9-a2e2-9916e26ef920\" (UID: \"da93db43-37fc-40c9-a2e2-9916e26ef920\") " Nov 23 07:09:22 crc kubenswrapper[4559]: I1123 07:09:22.961005 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da93db43-37fc-40c9-a2e2-9916e26ef920-utilities\") pod \"da93db43-37fc-40c9-a2e2-9916e26ef920\" (UID: \"da93db43-37fc-40c9-a2e2-9916e26ef920\") " Nov 23 07:09:22 crc kubenswrapper[4559]: I1123 07:09:22.961693 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da93db43-37fc-40c9-a2e2-9916e26ef920-utilities" (OuterVolumeSpecName: "utilities") pod "da93db43-37fc-40c9-a2e2-9916e26ef920" (UID: "da93db43-37fc-40c9-a2e2-9916e26ef920"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:09:22 crc kubenswrapper[4559]: I1123 07:09:22.965416 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da93db43-37fc-40c9-a2e2-9916e26ef920-kube-api-access-2s5dc" (OuterVolumeSpecName: "kube-api-access-2s5dc") pod "da93db43-37fc-40c9-a2e2-9916e26ef920" (UID: "da93db43-37fc-40c9-a2e2-9916e26ef920"). InnerVolumeSpecName "kube-api-access-2s5dc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:09:22 crc kubenswrapper[4559]: I1123 07:09:22.999669 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da93db43-37fc-40c9-a2e2-9916e26ef920-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "da93db43-37fc-40c9-a2e2-9916e26ef920" (UID: "da93db43-37fc-40c9-a2e2-9916e26ef920"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:09:23 crc kubenswrapper[4559]: I1123 07:09:23.063008 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da93db43-37fc-40c9-a2e2-9916e26ef920-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:23 crc kubenswrapper[4559]: I1123 07:09:23.063035 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2s5dc\" (UniqueName: \"kubernetes.io/projected/da93db43-37fc-40c9-a2e2-9916e26ef920-kube-api-access-2s5dc\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:23 crc kubenswrapper[4559]: I1123 07:09:23.063046 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da93db43-37fc-40c9-a2e2-9916e26ef920-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:23 crc kubenswrapper[4559]: I1123 07:09:23.533555 4559 generic.go:334] "Generic (PLEG): container finished" podID="45c516df-85dd-4a43-94ef-ea1381622c06" containerID="e49821d0092fdae1e63db0c0a13a10e0a8843316db83f6e7b812cc247a31aa44" exitCode=0 Nov 23 07:09:23 crc kubenswrapper[4559]: I1123 07:09:23.533622 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5" event={"ID":"45c516df-85dd-4a43-94ef-ea1381622c06","Type":"ContainerDied","Data":"e49821d0092fdae1e63db0c0a13a10e0a8843316db83f6e7b812cc247a31aa44"} Nov 23 07:09:23 crc kubenswrapper[4559]: I1123 07:09:23.535834 4559 generic.go:334] "Generic (PLEG): container finished" podID="da93db43-37fc-40c9-a2e2-9916e26ef920" containerID="ad819b987952cdf194b260e62c60987225a8a7d2e40084f66250e41664b1806d" exitCode=0 Nov 23 07:09:23 crc kubenswrapper[4559]: I1123 07:09:23.535880 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hhqzb" event={"ID":"da93db43-37fc-40c9-a2e2-9916e26ef920","Type":"ContainerDied","Data":"ad819b987952cdf194b260e62c60987225a8a7d2e40084f66250e41664b1806d"} Nov 23 07:09:23 crc kubenswrapper[4559]: I1123 07:09:23.535892 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hhqzb" Nov 23 07:09:23 crc kubenswrapper[4559]: I1123 07:09:23.535913 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hhqzb" event={"ID":"da93db43-37fc-40c9-a2e2-9916e26ef920","Type":"ContainerDied","Data":"ea59ceea00fb775adc30d702bfed181d19eaa2bed2339b745075c6c0b874f260"} Nov 23 07:09:23 crc kubenswrapper[4559]: I1123 07:09:23.535933 4559 scope.go:117] "RemoveContainer" containerID="ad819b987952cdf194b260e62c60987225a8a7d2e40084f66250e41664b1806d" Nov 23 07:09:23 crc kubenswrapper[4559]: I1123 07:09:23.560216 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hhqzb"] Nov 23 07:09:23 crc kubenswrapper[4559]: I1123 07:09:23.565622 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hhqzb"] Nov 23 07:09:23 crc kubenswrapper[4559]: I1123 07:09:23.566186 4559 scope.go:117] "RemoveContainer" containerID="2e4572001f1666e734759d0badef211fe79d01c9e82dc39425c538b038a9474e" Nov 23 07:09:23 crc kubenswrapper[4559]: I1123 07:09:23.597061 4559 scope.go:117] "RemoveContainer" containerID="b2766e8e6f8c3caad942345a56a48163ac88d207a21f42e35c1ab1902004e1b0" Nov 23 07:09:23 crc kubenswrapper[4559]: I1123 07:09:23.612249 4559 scope.go:117] "RemoveContainer" containerID="ad819b987952cdf194b260e62c60987225a8a7d2e40084f66250e41664b1806d" Nov 23 07:09:23 crc kubenswrapper[4559]: E1123 07:09:23.612684 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad819b987952cdf194b260e62c60987225a8a7d2e40084f66250e41664b1806d\": container with ID starting with ad819b987952cdf194b260e62c60987225a8a7d2e40084f66250e41664b1806d not found: ID does not exist" containerID="ad819b987952cdf194b260e62c60987225a8a7d2e40084f66250e41664b1806d" Nov 23 07:09:23 crc kubenswrapper[4559]: I1123 07:09:23.612718 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad819b987952cdf194b260e62c60987225a8a7d2e40084f66250e41664b1806d"} err="failed to get container status \"ad819b987952cdf194b260e62c60987225a8a7d2e40084f66250e41664b1806d\": rpc error: code = NotFound desc = could not find container \"ad819b987952cdf194b260e62c60987225a8a7d2e40084f66250e41664b1806d\": container with ID starting with ad819b987952cdf194b260e62c60987225a8a7d2e40084f66250e41664b1806d not found: ID does not exist" Nov 23 07:09:23 crc kubenswrapper[4559]: I1123 07:09:23.612740 4559 scope.go:117] "RemoveContainer" containerID="2e4572001f1666e734759d0badef211fe79d01c9e82dc39425c538b038a9474e" Nov 23 07:09:23 crc kubenswrapper[4559]: E1123 07:09:23.613093 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e4572001f1666e734759d0badef211fe79d01c9e82dc39425c538b038a9474e\": container with ID starting with 2e4572001f1666e734759d0badef211fe79d01c9e82dc39425c538b038a9474e not found: ID does not exist" containerID="2e4572001f1666e734759d0badef211fe79d01c9e82dc39425c538b038a9474e" Nov 23 07:09:23 crc kubenswrapper[4559]: I1123 07:09:23.613129 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e4572001f1666e734759d0badef211fe79d01c9e82dc39425c538b038a9474e"} err="failed to get container status \"2e4572001f1666e734759d0badef211fe79d01c9e82dc39425c538b038a9474e\": rpc error: code = NotFound desc = could not find container \"2e4572001f1666e734759d0badef211fe79d01c9e82dc39425c538b038a9474e\": container with ID starting with 2e4572001f1666e734759d0badef211fe79d01c9e82dc39425c538b038a9474e not found: ID does not exist" Nov 23 07:09:23 crc kubenswrapper[4559]: I1123 07:09:23.613152 4559 scope.go:117] "RemoveContainer" containerID="b2766e8e6f8c3caad942345a56a48163ac88d207a21f42e35c1ab1902004e1b0" Nov 23 07:09:23 crc kubenswrapper[4559]: E1123 07:09:23.613384 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2766e8e6f8c3caad942345a56a48163ac88d207a21f42e35c1ab1902004e1b0\": container with ID starting with b2766e8e6f8c3caad942345a56a48163ac88d207a21f42e35c1ab1902004e1b0 not found: ID does not exist" containerID="b2766e8e6f8c3caad942345a56a48163ac88d207a21f42e35c1ab1902004e1b0" Nov 23 07:09:23 crc kubenswrapper[4559]: I1123 07:09:23.613481 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2766e8e6f8c3caad942345a56a48163ac88d207a21f42e35c1ab1902004e1b0"} err="failed to get container status \"b2766e8e6f8c3caad942345a56a48163ac88d207a21f42e35c1ab1902004e1b0\": rpc error: code = NotFound desc = could not find container \"b2766e8e6f8c3caad942345a56a48163ac88d207a21f42e35c1ab1902004e1b0\": container with ID starting with b2766e8e6f8c3caad942345a56a48163ac88d207a21f42e35c1ab1902004e1b0 not found: ID does not exist" Nov 23 07:09:24 crc kubenswrapper[4559]: I1123 07:09:24.028222 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-b6wwk"] Nov 23 07:09:24 crc kubenswrapper[4559]: I1123 07:09:24.033792 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-b6wwk"] Nov 23 07:09:24 crc kubenswrapper[4559]: I1123 07:09:24.281634 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da93db43-37fc-40c9-a2e2-9916e26ef920" path="/var/lib/kubelet/pods/da93db43-37fc-40c9-a2e2-9916e26ef920/volumes" Nov 23 07:09:24 crc kubenswrapper[4559]: I1123 07:09:24.282372 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e43dc009-537b-4a2c-b481-e42f41eec38d" path="/var/lib/kubelet/pods/e43dc009-537b-4a2c-b481-e42f41eec38d/volumes" Nov 23 07:09:24 crc kubenswrapper[4559]: I1123 07:09:24.819612 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5" Nov 23 07:09:24 crc kubenswrapper[4559]: I1123 07:09:24.889531 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/45c516df-85dd-4a43-94ef-ea1381622c06-ssh-key\") pod \"45c516df-85dd-4a43-94ef-ea1381622c06\" (UID: \"45c516df-85dd-4a43-94ef-ea1381622c06\") " Nov 23 07:09:24 crc kubenswrapper[4559]: I1123 07:09:24.889612 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/45c516df-85dd-4a43-94ef-ea1381622c06-inventory\") pod \"45c516df-85dd-4a43-94ef-ea1381622c06\" (UID: \"45c516df-85dd-4a43-94ef-ea1381622c06\") " Nov 23 07:09:24 crc kubenswrapper[4559]: I1123 07:09:24.889635 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8ncb\" (UniqueName: \"kubernetes.io/projected/45c516df-85dd-4a43-94ef-ea1381622c06-kube-api-access-m8ncb\") pod \"45c516df-85dd-4a43-94ef-ea1381622c06\" (UID: \"45c516df-85dd-4a43-94ef-ea1381622c06\") " Nov 23 07:09:24 crc kubenswrapper[4559]: I1123 07:09:24.906752 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45c516df-85dd-4a43-94ef-ea1381622c06-kube-api-access-m8ncb" (OuterVolumeSpecName: "kube-api-access-m8ncb") pod "45c516df-85dd-4a43-94ef-ea1381622c06" (UID: "45c516df-85dd-4a43-94ef-ea1381622c06"). InnerVolumeSpecName "kube-api-access-m8ncb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:09:24 crc kubenswrapper[4559]: I1123 07:09:24.910496 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45c516df-85dd-4a43-94ef-ea1381622c06-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "45c516df-85dd-4a43-94ef-ea1381622c06" (UID: "45c516df-85dd-4a43-94ef-ea1381622c06"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:09:24 crc kubenswrapper[4559]: I1123 07:09:24.912010 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45c516df-85dd-4a43-94ef-ea1381622c06-inventory" (OuterVolumeSpecName: "inventory") pod "45c516df-85dd-4a43-94ef-ea1381622c06" (UID: "45c516df-85dd-4a43-94ef-ea1381622c06"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:09:24 crc kubenswrapper[4559]: I1123 07:09:24.991604 4559 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/45c516df-85dd-4a43-94ef-ea1381622c06-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:24 crc kubenswrapper[4559]: I1123 07:09:24.991628 4559 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/45c516df-85dd-4a43-94ef-ea1381622c06-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:24 crc kubenswrapper[4559]: I1123 07:09:24.991653 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8ncb\" (UniqueName: \"kubernetes.io/projected/45c516df-85dd-4a43-94ef-ea1381622c06-kube-api-access-m8ncb\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:25 crc kubenswrapper[4559]: I1123 07:09:25.551329 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5" event={"ID":"45c516df-85dd-4a43-94ef-ea1381622c06","Type":"ContainerDied","Data":"d490683b13a70ef53e47b4a0208933c3b588f4a064aeb5968e6a68a834bae20a"} Nov 23 07:09:25 crc kubenswrapper[4559]: I1123 07:09:25.551529 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d490683b13a70ef53e47b4a0208933c3b588f4a064aeb5968e6a68a834bae20a" Nov 23 07:09:25 crc kubenswrapper[4559]: I1123 07:09:25.551376 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5" Nov 23 07:09:25 crc kubenswrapper[4559]: I1123 07:09:25.599132 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crmzz"] Nov 23 07:09:25 crc kubenswrapper[4559]: E1123 07:09:25.599450 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da93db43-37fc-40c9-a2e2-9916e26ef920" containerName="extract-utilities" Nov 23 07:09:25 crc kubenswrapper[4559]: I1123 07:09:25.599469 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="da93db43-37fc-40c9-a2e2-9916e26ef920" containerName="extract-utilities" Nov 23 07:09:25 crc kubenswrapper[4559]: E1123 07:09:25.599478 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da93db43-37fc-40c9-a2e2-9916e26ef920" containerName="registry-server" Nov 23 07:09:25 crc kubenswrapper[4559]: I1123 07:09:25.599484 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="da93db43-37fc-40c9-a2e2-9916e26ef920" containerName="registry-server" Nov 23 07:09:25 crc kubenswrapper[4559]: E1123 07:09:25.599502 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45c516df-85dd-4a43-94ef-ea1381622c06" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 23 07:09:25 crc kubenswrapper[4559]: I1123 07:09:25.599509 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="45c516df-85dd-4a43-94ef-ea1381622c06" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 23 07:09:25 crc kubenswrapper[4559]: E1123 07:09:25.599527 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da93db43-37fc-40c9-a2e2-9916e26ef920" containerName="extract-content" Nov 23 07:09:25 crc kubenswrapper[4559]: I1123 07:09:25.599534 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="da93db43-37fc-40c9-a2e2-9916e26ef920" containerName="extract-content" Nov 23 07:09:25 crc kubenswrapper[4559]: I1123 07:09:25.599725 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="da93db43-37fc-40c9-a2e2-9916e26ef920" containerName="registry-server" Nov 23 07:09:25 crc kubenswrapper[4559]: I1123 07:09:25.599755 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="45c516df-85dd-4a43-94ef-ea1381622c06" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 23 07:09:25 crc kubenswrapper[4559]: I1123 07:09:25.600280 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crmzz" Nov 23 07:09:25 crc kubenswrapper[4559]: I1123 07:09:25.602149 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 07:09:25 crc kubenswrapper[4559]: I1123 07:09:25.602201 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 07:09:25 crc kubenswrapper[4559]: I1123 07:09:25.602314 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mj6c" Nov 23 07:09:25 crc kubenswrapper[4559]: I1123 07:09:25.606046 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 07:09:25 crc kubenswrapper[4559]: I1123 07:09:25.612186 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crmzz"] Nov 23 07:09:25 crc kubenswrapper[4559]: I1123 07:09:25.702112 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8fb6\" (UniqueName: \"kubernetes.io/projected/2368491a-9b45-4c74-bb8d-b6be5fca89f8-kube-api-access-m8fb6\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-crmzz\" (UID: \"2368491a-9b45-4c74-bb8d-b6be5fca89f8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crmzz" Nov 23 07:09:25 crc kubenswrapper[4559]: I1123 07:09:25.702423 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2368491a-9b45-4c74-bb8d-b6be5fca89f8-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-crmzz\" (UID: \"2368491a-9b45-4c74-bb8d-b6be5fca89f8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crmzz" Nov 23 07:09:25 crc kubenswrapper[4559]: I1123 07:09:25.702550 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2368491a-9b45-4c74-bb8d-b6be5fca89f8-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-crmzz\" (UID: \"2368491a-9b45-4c74-bb8d-b6be5fca89f8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crmzz" Nov 23 07:09:25 crc kubenswrapper[4559]: I1123 07:09:25.803633 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2368491a-9b45-4c74-bb8d-b6be5fca89f8-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-crmzz\" (UID: \"2368491a-9b45-4c74-bb8d-b6be5fca89f8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crmzz" Nov 23 07:09:25 crc kubenswrapper[4559]: I1123 07:09:25.803735 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2368491a-9b45-4c74-bb8d-b6be5fca89f8-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-crmzz\" (UID: \"2368491a-9b45-4c74-bb8d-b6be5fca89f8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crmzz" Nov 23 07:09:25 crc kubenswrapper[4559]: I1123 07:09:25.803796 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8fb6\" (UniqueName: \"kubernetes.io/projected/2368491a-9b45-4c74-bb8d-b6be5fca89f8-kube-api-access-m8fb6\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-crmzz\" (UID: \"2368491a-9b45-4c74-bb8d-b6be5fca89f8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crmzz" Nov 23 07:09:25 crc kubenswrapper[4559]: I1123 07:09:25.806685 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2368491a-9b45-4c74-bb8d-b6be5fca89f8-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-crmzz\" (UID: \"2368491a-9b45-4c74-bb8d-b6be5fca89f8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crmzz" Nov 23 07:09:25 crc kubenswrapper[4559]: I1123 07:09:25.806771 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2368491a-9b45-4c74-bb8d-b6be5fca89f8-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-crmzz\" (UID: \"2368491a-9b45-4c74-bb8d-b6be5fca89f8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crmzz" Nov 23 07:09:25 crc kubenswrapper[4559]: I1123 07:09:25.818843 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8fb6\" (UniqueName: \"kubernetes.io/projected/2368491a-9b45-4c74-bb8d-b6be5fca89f8-kube-api-access-m8fb6\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-crmzz\" (UID: \"2368491a-9b45-4c74-bb8d-b6be5fca89f8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crmzz" Nov 23 07:09:25 crc kubenswrapper[4559]: I1123 07:09:25.913767 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crmzz" Nov 23 07:09:26 crc kubenswrapper[4559]: I1123 07:09:26.167195 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:09:26 crc kubenswrapper[4559]: I1123 07:09:26.167246 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:09:26 crc kubenswrapper[4559]: I1123 07:09:26.167285 4559 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" Nov 23 07:09:26 crc kubenswrapper[4559]: I1123 07:09:26.167915 4559 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b"} pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 07:09:26 crc kubenswrapper[4559]: I1123 07:09:26.167978 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" containerID="cri-o://ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b" gracePeriod=600 Nov 23 07:09:26 crc kubenswrapper[4559]: E1123 07:09:26.288393 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:09:26 crc kubenswrapper[4559]: I1123 07:09:26.324289 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crmzz"] Nov 23 07:09:26 crc kubenswrapper[4559]: I1123 07:09:26.559824 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crmzz" event={"ID":"2368491a-9b45-4c74-bb8d-b6be5fca89f8","Type":"ContainerStarted","Data":"83b10d1fb52b038c7f07e57d573402c9c18f8b3d17011141676ec8fb3637265a"} Nov 23 07:09:26 crc kubenswrapper[4559]: I1123 07:09:26.561962 4559 generic.go:334] "Generic (PLEG): container finished" podID="4731beee-0cac-4189-8a70-743b0b709095" containerID="ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b" exitCode=0 Nov 23 07:09:26 crc kubenswrapper[4559]: I1123 07:09:26.561999 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerDied","Data":"ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b"} Nov 23 07:09:26 crc kubenswrapper[4559]: I1123 07:09:26.562034 4559 scope.go:117] "RemoveContainer" containerID="7f8ad51d7a50444a4cc31965b149316d371db5ed9346214260de65924e42fa55" Nov 23 07:09:26 crc kubenswrapper[4559]: I1123 07:09:26.562602 4559 scope.go:117] "RemoveContainer" containerID="ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b" Nov 23 07:09:26 crc kubenswrapper[4559]: E1123 07:09:26.562970 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:09:27 crc kubenswrapper[4559]: I1123 07:09:27.569738 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crmzz" event={"ID":"2368491a-9b45-4c74-bb8d-b6be5fca89f8","Type":"ContainerStarted","Data":"a856672418a852a7edd51e4a5e432598e8a8d4d710e58fb82809e51dddd90393"} Nov 23 07:09:27 crc kubenswrapper[4559]: I1123 07:09:27.584334 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crmzz" podStartSLOduration=2.012608382 podStartE2EDuration="2.584318298s" podCreationTimestamp="2025-11-23 07:09:25 +0000 UTC" firstStartedPulling="2025-11-23 07:09:26.329490334 +0000 UTC m=+1468.351475949" lastFinishedPulling="2025-11-23 07:09:26.901200251 +0000 UTC m=+1468.923185865" observedRunningTime="2025-11-23 07:09:27.579809358 +0000 UTC m=+1469.601794972" watchObservedRunningTime="2025-11-23 07:09:27.584318298 +0000 UTC m=+1469.606303912" Nov 23 07:09:30 crc kubenswrapper[4559]: I1123 07:09:30.592761 4559 generic.go:334] "Generic (PLEG): container finished" podID="2368491a-9b45-4c74-bb8d-b6be5fca89f8" containerID="a856672418a852a7edd51e4a5e432598e8a8d4d710e58fb82809e51dddd90393" exitCode=0 Nov 23 07:09:30 crc kubenswrapper[4559]: I1123 07:09:30.592848 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crmzz" event={"ID":"2368491a-9b45-4c74-bb8d-b6be5fca89f8","Type":"ContainerDied","Data":"a856672418a852a7edd51e4a5e432598e8a8d4d710e58fb82809e51dddd90393"} Nov 23 07:09:31 crc kubenswrapper[4559]: I1123 07:09:31.885694 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crmzz" Nov 23 07:09:31 crc kubenswrapper[4559]: I1123 07:09:31.996497 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2368491a-9b45-4c74-bb8d-b6be5fca89f8-inventory\") pod \"2368491a-9b45-4c74-bb8d-b6be5fca89f8\" (UID: \"2368491a-9b45-4c74-bb8d-b6be5fca89f8\") " Nov 23 07:09:31 crc kubenswrapper[4559]: I1123 07:09:31.996574 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2368491a-9b45-4c74-bb8d-b6be5fca89f8-ssh-key\") pod \"2368491a-9b45-4c74-bb8d-b6be5fca89f8\" (UID: \"2368491a-9b45-4c74-bb8d-b6be5fca89f8\") " Nov 23 07:09:31 crc kubenswrapper[4559]: I1123 07:09:31.996661 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8fb6\" (UniqueName: \"kubernetes.io/projected/2368491a-9b45-4c74-bb8d-b6be5fca89f8-kube-api-access-m8fb6\") pod \"2368491a-9b45-4c74-bb8d-b6be5fca89f8\" (UID: \"2368491a-9b45-4c74-bb8d-b6be5fca89f8\") " Nov 23 07:09:32 crc kubenswrapper[4559]: I1123 07:09:32.000713 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2368491a-9b45-4c74-bb8d-b6be5fca89f8-kube-api-access-m8fb6" (OuterVolumeSpecName: "kube-api-access-m8fb6") pod "2368491a-9b45-4c74-bb8d-b6be5fca89f8" (UID: "2368491a-9b45-4c74-bb8d-b6be5fca89f8"). InnerVolumeSpecName "kube-api-access-m8fb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:09:32 crc kubenswrapper[4559]: I1123 07:09:32.018043 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2368491a-9b45-4c74-bb8d-b6be5fca89f8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2368491a-9b45-4c74-bb8d-b6be5fca89f8" (UID: "2368491a-9b45-4c74-bb8d-b6be5fca89f8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:09:32 crc kubenswrapper[4559]: I1123 07:09:32.018602 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2368491a-9b45-4c74-bb8d-b6be5fca89f8-inventory" (OuterVolumeSpecName: "inventory") pod "2368491a-9b45-4c74-bb8d-b6be5fca89f8" (UID: "2368491a-9b45-4c74-bb8d-b6be5fca89f8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:09:32 crc kubenswrapper[4559]: I1123 07:09:32.098210 4559 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2368491a-9b45-4c74-bb8d-b6be5fca89f8-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:32 crc kubenswrapper[4559]: I1123 07:09:32.098241 4559 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2368491a-9b45-4c74-bb8d-b6be5fca89f8-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:32 crc kubenswrapper[4559]: I1123 07:09:32.098252 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8fb6\" (UniqueName: \"kubernetes.io/projected/2368491a-9b45-4c74-bb8d-b6be5fca89f8-kube-api-access-m8fb6\") on node \"crc\" DevicePath \"\"" Nov 23 07:09:32 crc kubenswrapper[4559]: I1123 07:09:32.607169 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crmzz" event={"ID":"2368491a-9b45-4c74-bb8d-b6be5fca89f8","Type":"ContainerDied","Data":"83b10d1fb52b038c7f07e57d573402c9c18f8b3d17011141676ec8fb3637265a"} Nov 23 07:09:32 crc kubenswrapper[4559]: I1123 07:09:32.607204 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83b10d1fb52b038c7f07e57d573402c9c18f8b3d17011141676ec8fb3637265a" Nov 23 07:09:32 crc kubenswrapper[4559]: I1123 07:09:32.607218 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-crmzz" Nov 23 07:09:32 crc kubenswrapper[4559]: I1123 07:09:32.654981 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-wgnh6"] Nov 23 07:09:32 crc kubenswrapper[4559]: E1123 07:09:32.655313 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2368491a-9b45-4c74-bb8d-b6be5fca89f8" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 23 07:09:32 crc kubenswrapper[4559]: I1123 07:09:32.655332 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="2368491a-9b45-4c74-bb8d-b6be5fca89f8" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 23 07:09:32 crc kubenswrapper[4559]: I1123 07:09:32.655511 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="2368491a-9b45-4c74-bb8d-b6be5fca89f8" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 23 07:09:32 crc kubenswrapper[4559]: I1123 07:09:32.656063 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wgnh6" Nov 23 07:09:32 crc kubenswrapper[4559]: I1123 07:09:32.657874 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 07:09:32 crc kubenswrapper[4559]: I1123 07:09:32.663957 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 07:09:32 crc kubenswrapper[4559]: I1123 07:09:32.664462 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mj6c" Nov 23 07:09:32 crc kubenswrapper[4559]: I1123 07:09:32.664495 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-wgnh6"] Nov 23 07:09:32 crc kubenswrapper[4559]: I1123 07:09:32.664562 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 07:09:32 crc kubenswrapper[4559]: I1123 07:09:32.810004 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/460a9b32-addb-458a-9913-4ba0433b3eb3-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-wgnh6\" (UID: \"460a9b32-addb-458a-9913-4ba0433b3eb3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wgnh6" Nov 23 07:09:32 crc kubenswrapper[4559]: I1123 07:09:32.810086 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2p8n\" (UniqueName: \"kubernetes.io/projected/460a9b32-addb-458a-9913-4ba0433b3eb3-kube-api-access-s2p8n\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-wgnh6\" (UID: \"460a9b32-addb-458a-9913-4ba0433b3eb3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wgnh6" Nov 23 07:09:32 crc kubenswrapper[4559]: I1123 07:09:32.810119 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/460a9b32-addb-458a-9913-4ba0433b3eb3-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-wgnh6\" (UID: \"460a9b32-addb-458a-9913-4ba0433b3eb3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wgnh6" Nov 23 07:09:32 crc kubenswrapper[4559]: I1123 07:09:32.912009 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/460a9b32-addb-458a-9913-4ba0433b3eb3-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-wgnh6\" (UID: \"460a9b32-addb-458a-9913-4ba0433b3eb3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wgnh6" Nov 23 07:09:32 crc kubenswrapper[4559]: I1123 07:09:32.912086 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2p8n\" (UniqueName: \"kubernetes.io/projected/460a9b32-addb-458a-9913-4ba0433b3eb3-kube-api-access-s2p8n\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-wgnh6\" (UID: \"460a9b32-addb-458a-9913-4ba0433b3eb3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wgnh6" Nov 23 07:09:32 crc kubenswrapper[4559]: I1123 07:09:32.912116 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/460a9b32-addb-458a-9913-4ba0433b3eb3-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-wgnh6\" (UID: \"460a9b32-addb-458a-9913-4ba0433b3eb3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wgnh6" Nov 23 07:09:32 crc kubenswrapper[4559]: I1123 07:09:32.916991 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/460a9b32-addb-458a-9913-4ba0433b3eb3-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-wgnh6\" (UID: \"460a9b32-addb-458a-9913-4ba0433b3eb3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wgnh6" Nov 23 07:09:32 crc kubenswrapper[4559]: I1123 07:09:32.917876 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/460a9b32-addb-458a-9913-4ba0433b3eb3-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-wgnh6\" (UID: \"460a9b32-addb-458a-9913-4ba0433b3eb3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wgnh6" Nov 23 07:09:32 crc kubenswrapper[4559]: I1123 07:09:32.925540 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2p8n\" (UniqueName: \"kubernetes.io/projected/460a9b32-addb-458a-9913-4ba0433b3eb3-kube-api-access-s2p8n\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-wgnh6\" (UID: \"460a9b32-addb-458a-9913-4ba0433b3eb3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wgnh6" Nov 23 07:09:32 crc kubenswrapper[4559]: I1123 07:09:32.969325 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wgnh6" Nov 23 07:09:33 crc kubenswrapper[4559]: I1123 07:09:33.396993 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-wgnh6"] Nov 23 07:09:33 crc kubenswrapper[4559]: W1123 07:09:33.401063 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod460a9b32_addb_458a_9913_4ba0433b3eb3.slice/crio-b7fa448ce36e4b7673a1c755cd1f0b9fae80d42b1e06716dd77c65f4433398fa WatchSource:0}: Error finding container b7fa448ce36e4b7673a1c755cd1f0b9fae80d42b1e06716dd77c65f4433398fa: Status 404 returned error can't find the container with id b7fa448ce36e4b7673a1c755cd1f0b9fae80d42b1e06716dd77c65f4433398fa Nov 23 07:09:33 crc kubenswrapper[4559]: I1123 07:09:33.614631 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wgnh6" event={"ID":"460a9b32-addb-458a-9913-4ba0433b3eb3","Type":"ContainerStarted","Data":"b7fa448ce36e4b7673a1c755cd1f0b9fae80d42b1e06716dd77c65f4433398fa"} Nov 23 07:09:34 crc kubenswrapper[4559]: I1123 07:09:34.622121 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wgnh6" event={"ID":"460a9b32-addb-458a-9913-4ba0433b3eb3","Type":"ContainerStarted","Data":"16015a88fd0d4a93c657fe32736088cf9552d194c01fb2a7416c77d4e3c1d187"} Nov 23 07:09:34 crc kubenswrapper[4559]: I1123 07:09:34.637666 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wgnh6" podStartSLOduration=2.173829547 podStartE2EDuration="2.63763745s" podCreationTimestamp="2025-11-23 07:09:32 +0000 UTC" firstStartedPulling="2025-11-23 07:09:33.403038268 +0000 UTC m=+1475.425023882" lastFinishedPulling="2025-11-23 07:09:33.866846171 +0000 UTC m=+1475.888831785" observedRunningTime="2025-11-23 07:09:34.633029343 +0000 UTC m=+1476.655014956" watchObservedRunningTime="2025-11-23 07:09:34.63763745 +0000 UTC m=+1476.659623064" Nov 23 07:09:36 crc kubenswrapper[4559]: I1123 07:09:36.020897 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-qgn9k"] Nov 23 07:09:36 crc kubenswrapper[4559]: I1123 07:09:36.027347 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-qgn9k"] Nov 23 07:09:36 crc kubenswrapper[4559]: I1123 07:09:36.282681 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad2af1c5-49df-4093-a38f-b1b47e2ddc1f" path="/var/lib/kubelet/pods/ad2af1c5-49df-4093-a38f-b1b47e2ddc1f/volumes" Nov 23 07:09:39 crc kubenswrapper[4559]: I1123 07:09:39.273673 4559 scope.go:117] "RemoveContainer" containerID="ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b" Nov 23 07:09:39 crc kubenswrapper[4559]: E1123 07:09:39.274839 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:09:40 crc kubenswrapper[4559]: I1123 07:09:40.030367 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-mvljp"] Nov 23 07:09:40 crc kubenswrapper[4559]: I1123 07:09:40.037565 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-mvljp"] Nov 23 07:09:40 crc kubenswrapper[4559]: I1123 07:09:40.283327 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43aed302-5f95-4e51-b845-5a9cee9d1a08" path="/var/lib/kubelet/pods/43aed302-5f95-4e51-b845-5a9cee9d1a08/volumes" Nov 23 07:09:49 crc kubenswrapper[4559]: I1123 07:09:49.020018 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-k6zjg"] Nov 23 07:09:49 crc kubenswrapper[4559]: I1123 07:09:49.025605 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-k6zjg"] Nov 23 07:09:49 crc kubenswrapper[4559]: I1123 07:09:49.031381 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-hz99c"] Nov 23 07:09:49 crc kubenswrapper[4559]: I1123 07:09:49.036212 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-hz99c"] Nov 23 07:09:50 crc kubenswrapper[4559]: I1123 07:09:50.273302 4559 scope.go:117] "RemoveContainer" containerID="ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b" Nov 23 07:09:50 crc kubenswrapper[4559]: E1123 07:09:50.273732 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:09:50 crc kubenswrapper[4559]: I1123 07:09:50.283944 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0534e805-e3fd-4193-80ed-3c9fa69a041a" path="/var/lib/kubelet/pods/0534e805-e3fd-4193-80ed-3c9fa69a041a/volumes" Nov 23 07:09:50 crc kubenswrapper[4559]: I1123 07:09:50.284665 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b14d2018-932c-47b6-bba9-232dff3ce320" path="/var/lib/kubelet/pods/b14d2018-932c-47b6-bba9-232dff3ce320/volumes" Nov 23 07:09:59 crc kubenswrapper[4559]: I1123 07:09:59.032602 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-bxpjp"] Nov 23 07:09:59 crc kubenswrapper[4559]: I1123 07:09:59.038484 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-bxpjp"] Nov 23 07:09:59 crc kubenswrapper[4559]: I1123 07:09:59.792882 4559 generic.go:334] "Generic (PLEG): container finished" podID="460a9b32-addb-458a-9913-4ba0433b3eb3" containerID="16015a88fd0d4a93c657fe32736088cf9552d194c01fb2a7416c77d4e3c1d187" exitCode=0 Nov 23 07:09:59 crc kubenswrapper[4559]: I1123 07:09:59.792961 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wgnh6" event={"ID":"460a9b32-addb-458a-9913-4ba0433b3eb3","Type":"ContainerDied","Data":"16015a88fd0d4a93c657fe32736088cf9552d194c01fb2a7416c77d4e3c1d187"} Nov 23 07:10:00 crc kubenswrapper[4559]: I1123 07:10:00.281290 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8ae1721-4ded-4e7d-9091-88ac908d0554" path="/var/lib/kubelet/pods/c8ae1721-4ded-4e7d-9091-88ac908d0554/volumes" Nov 23 07:10:01 crc kubenswrapper[4559]: I1123 07:10:01.100997 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wgnh6" Nov 23 07:10:01 crc kubenswrapper[4559]: I1123 07:10:01.247225 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/460a9b32-addb-458a-9913-4ba0433b3eb3-inventory\") pod \"460a9b32-addb-458a-9913-4ba0433b3eb3\" (UID: \"460a9b32-addb-458a-9913-4ba0433b3eb3\") " Nov 23 07:10:01 crc kubenswrapper[4559]: I1123 07:10:01.247267 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/460a9b32-addb-458a-9913-4ba0433b3eb3-ssh-key\") pod \"460a9b32-addb-458a-9913-4ba0433b3eb3\" (UID: \"460a9b32-addb-458a-9913-4ba0433b3eb3\") " Nov 23 07:10:01 crc kubenswrapper[4559]: I1123 07:10:01.247314 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2p8n\" (UniqueName: \"kubernetes.io/projected/460a9b32-addb-458a-9913-4ba0433b3eb3-kube-api-access-s2p8n\") pod \"460a9b32-addb-458a-9913-4ba0433b3eb3\" (UID: \"460a9b32-addb-458a-9913-4ba0433b3eb3\") " Nov 23 07:10:01 crc kubenswrapper[4559]: I1123 07:10:01.251300 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/460a9b32-addb-458a-9913-4ba0433b3eb3-kube-api-access-s2p8n" (OuterVolumeSpecName: "kube-api-access-s2p8n") pod "460a9b32-addb-458a-9913-4ba0433b3eb3" (UID: "460a9b32-addb-458a-9913-4ba0433b3eb3"). InnerVolumeSpecName "kube-api-access-s2p8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:10:01 crc kubenswrapper[4559]: E1123 07:10:01.265867 4559 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/460a9b32-addb-458a-9913-4ba0433b3eb3-inventory podName:460a9b32-addb-458a-9913-4ba0433b3eb3 nodeName:}" failed. No retries permitted until 2025-11-23 07:10:01.76584705 +0000 UTC m=+1503.787832665 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "inventory" (UniqueName: "kubernetes.io/secret/460a9b32-addb-458a-9913-4ba0433b3eb3-inventory") pod "460a9b32-addb-458a-9913-4ba0433b3eb3" (UID: "460a9b32-addb-458a-9913-4ba0433b3eb3") : error deleting /var/lib/kubelet/pods/460a9b32-addb-458a-9913-4ba0433b3eb3/volume-subpaths: remove /var/lib/kubelet/pods/460a9b32-addb-458a-9913-4ba0433b3eb3/volume-subpaths: no such file or directory Nov 23 07:10:01 crc kubenswrapper[4559]: I1123 07:10:01.267633 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/460a9b32-addb-458a-9913-4ba0433b3eb3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "460a9b32-addb-458a-9913-4ba0433b3eb3" (UID: "460a9b32-addb-458a-9913-4ba0433b3eb3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:01 crc kubenswrapper[4559]: I1123 07:10:01.273256 4559 scope.go:117] "RemoveContainer" containerID="ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b" Nov 23 07:10:01 crc kubenswrapper[4559]: E1123 07:10:01.273561 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:10:01 crc kubenswrapper[4559]: I1123 07:10:01.349184 4559 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/460a9b32-addb-458a-9913-4ba0433b3eb3-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:01 crc kubenswrapper[4559]: I1123 07:10:01.349209 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2p8n\" (UniqueName: \"kubernetes.io/projected/460a9b32-addb-458a-9913-4ba0433b3eb3-kube-api-access-s2p8n\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:01 crc kubenswrapper[4559]: I1123 07:10:01.807178 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wgnh6" event={"ID":"460a9b32-addb-458a-9913-4ba0433b3eb3","Type":"ContainerDied","Data":"b7fa448ce36e4b7673a1c755cd1f0b9fae80d42b1e06716dd77c65f4433398fa"} Nov 23 07:10:01 crc kubenswrapper[4559]: I1123 07:10:01.807215 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7fa448ce36e4b7673a1c755cd1f0b9fae80d42b1e06716dd77c65f4433398fa" Nov 23 07:10:01 crc kubenswrapper[4559]: I1123 07:10:01.807440 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wgnh6" Nov 23 07:10:01 crc kubenswrapper[4559]: I1123 07:10:01.856354 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/460a9b32-addb-458a-9913-4ba0433b3eb3-inventory\") pod \"460a9b32-addb-458a-9913-4ba0433b3eb3\" (UID: \"460a9b32-addb-458a-9913-4ba0433b3eb3\") " Nov 23 07:10:01 crc kubenswrapper[4559]: I1123 07:10:01.859801 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/460a9b32-addb-458a-9913-4ba0433b3eb3-inventory" (OuterVolumeSpecName: "inventory") pod "460a9b32-addb-458a-9913-4ba0433b3eb3" (UID: "460a9b32-addb-458a-9913-4ba0433b3eb3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:01 crc kubenswrapper[4559]: I1123 07:10:01.860888 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms"] Nov 23 07:10:01 crc kubenswrapper[4559]: E1123 07:10:01.861245 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="460a9b32-addb-458a-9913-4ba0433b3eb3" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 23 07:10:01 crc kubenswrapper[4559]: I1123 07:10:01.861264 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="460a9b32-addb-458a-9913-4ba0433b3eb3" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 23 07:10:01 crc kubenswrapper[4559]: I1123 07:10:01.861458 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="460a9b32-addb-458a-9913-4ba0433b3eb3" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 23 07:10:01 crc kubenswrapper[4559]: I1123 07:10:01.862076 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms" Nov 23 07:10:01 crc kubenswrapper[4559]: I1123 07:10:01.868215 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms"] Nov 23 07:10:01 crc kubenswrapper[4559]: I1123 07:10:01.959173 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ft2db\" (UniqueName: \"kubernetes.io/projected/f3cbccfd-3b9d-4d8d-aec1-702b2e494667-kube-api-access-ft2db\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms\" (UID: \"f3cbccfd-3b9d-4d8d-aec1-702b2e494667\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms" Nov 23 07:10:01 crc kubenswrapper[4559]: I1123 07:10:01.959249 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f3cbccfd-3b9d-4d8d-aec1-702b2e494667-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms\" (UID: \"f3cbccfd-3b9d-4d8d-aec1-702b2e494667\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms" Nov 23 07:10:01 crc kubenswrapper[4559]: I1123 07:10:01.959341 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3cbccfd-3b9d-4d8d-aec1-702b2e494667-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms\" (UID: \"f3cbccfd-3b9d-4d8d-aec1-702b2e494667\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms" Nov 23 07:10:01 crc kubenswrapper[4559]: I1123 07:10:01.959599 4559 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/460a9b32-addb-458a-9913-4ba0433b3eb3-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:02 crc kubenswrapper[4559]: I1123 07:10:02.061785 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ft2db\" (UniqueName: \"kubernetes.io/projected/f3cbccfd-3b9d-4d8d-aec1-702b2e494667-kube-api-access-ft2db\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms\" (UID: \"f3cbccfd-3b9d-4d8d-aec1-702b2e494667\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms" Nov 23 07:10:02 crc kubenswrapper[4559]: I1123 07:10:02.061859 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f3cbccfd-3b9d-4d8d-aec1-702b2e494667-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms\" (UID: \"f3cbccfd-3b9d-4d8d-aec1-702b2e494667\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms" Nov 23 07:10:02 crc kubenswrapper[4559]: I1123 07:10:02.061937 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3cbccfd-3b9d-4d8d-aec1-702b2e494667-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms\" (UID: \"f3cbccfd-3b9d-4d8d-aec1-702b2e494667\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms" Nov 23 07:10:02 crc kubenswrapper[4559]: I1123 07:10:02.065018 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f3cbccfd-3b9d-4d8d-aec1-702b2e494667-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms\" (UID: \"f3cbccfd-3b9d-4d8d-aec1-702b2e494667\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms" Nov 23 07:10:02 crc kubenswrapper[4559]: I1123 07:10:02.065355 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3cbccfd-3b9d-4d8d-aec1-702b2e494667-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms\" (UID: \"f3cbccfd-3b9d-4d8d-aec1-702b2e494667\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms" Nov 23 07:10:02 crc kubenswrapper[4559]: I1123 07:10:02.075910 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ft2db\" (UniqueName: \"kubernetes.io/projected/f3cbccfd-3b9d-4d8d-aec1-702b2e494667-kube-api-access-ft2db\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms\" (UID: \"f3cbccfd-3b9d-4d8d-aec1-702b2e494667\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms" Nov 23 07:10:02 crc kubenswrapper[4559]: I1123 07:10:02.211286 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms" Nov 23 07:10:02 crc kubenswrapper[4559]: I1123 07:10:02.644742 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms"] Nov 23 07:10:02 crc kubenswrapper[4559]: W1123 07:10:02.646703 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf3cbccfd_3b9d_4d8d_aec1_702b2e494667.slice/crio-83562b3814ebe0fc8529074af0d1ad41adca45cc34017cc6b6eaa16538d542b1 WatchSource:0}: Error finding container 83562b3814ebe0fc8529074af0d1ad41adca45cc34017cc6b6eaa16538d542b1: Status 404 returned error can't find the container with id 83562b3814ebe0fc8529074af0d1ad41adca45cc34017cc6b6eaa16538d542b1 Nov 23 07:10:02 crc kubenswrapper[4559]: I1123 07:10:02.815117 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms" event={"ID":"f3cbccfd-3b9d-4d8d-aec1-702b2e494667","Type":"ContainerStarted","Data":"83562b3814ebe0fc8529074af0d1ad41adca45cc34017cc6b6eaa16538d542b1"} Nov 23 07:10:03 crc kubenswrapper[4559]: I1123 07:10:03.824513 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms" event={"ID":"f3cbccfd-3b9d-4d8d-aec1-702b2e494667","Type":"ContainerStarted","Data":"07b908b4162cfb8b79b65c179b46ba62fe994ea5d7de6a2e8808a1de13f27098"} Nov 23 07:10:03 crc kubenswrapper[4559]: I1123 07:10:03.841039 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms" podStartSLOduration=2.336039817 podStartE2EDuration="2.841023203s" podCreationTimestamp="2025-11-23 07:10:01 +0000 UTC" firstStartedPulling="2025-11-23 07:10:02.648060168 +0000 UTC m=+1504.670045781" lastFinishedPulling="2025-11-23 07:10:03.153043553 +0000 UTC m=+1505.175029167" observedRunningTime="2025-11-23 07:10:03.835337752 +0000 UTC m=+1505.857323366" watchObservedRunningTime="2025-11-23 07:10:03.841023203 +0000 UTC m=+1505.863008817" Nov 23 07:10:12 crc kubenswrapper[4559]: I1123 07:10:12.672364 4559 scope.go:117] "RemoveContainer" containerID="006eea70d80fe9a828e3cf95c539095e2476f9d9af8cf91aaf392f78bea994e5" Nov 23 07:10:12 crc kubenswrapper[4559]: I1123 07:10:12.690576 4559 scope.go:117] "RemoveContainer" containerID="e2961ad63ff176f2f31a2c962fb7fbcdfa68d43edcc56527f0a2400a3e0ee978" Nov 23 07:10:12 crc kubenswrapper[4559]: I1123 07:10:12.724343 4559 scope.go:117] "RemoveContainer" containerID="8ba9f205a0ae7c46a2c64cf1fa99ff0dbb88b0dd9df79a738949e5d13c872fd7" Nov 23 07:10:12 crc kubenswrapper[4559]: I1123 07:10:12.759130 4559 scope.go:117] "RemoveContainer" containerID="112bc0f4b3c11d7acaad024495d518c9a4b4f11fca20992ff67b112146bba753" Nov 23 07:10:12 crc kubenswrapper[4559]: I1123 07:10:12.793666 4559 scope.go:117] "RemoveContainer" containerID="178d1c0f29794efd9880d96375c09387602578935cfa41dca7f541e19feea9ea" Nov 23 07:10:12 crc kubenswrapper[4559]: I1123 07:10:12.824848 4559 scope.go:117] "RemoveContainer" containerID="ffc5cf05f27d9f9aba701eab360263ff3bebe04d77619e8fa4aded2a481125eb" Nov 23 07:10:12 crc kubenswrapper[4559]: I1123 07:10:12.855956 4559 scope.go:117] "RemoveContainer" containerID="0631320823f51f57cba78a57d4205017d5ce7509f10c6009526c665185df09f3" Nov 23 07:10:12 crc kubenswrapper[4559]: I1123 07:10:12.872670 4559 scope.go:117] "RemoveContainer" containerID="69208340657fbd898448ec1513e3922261bc0f607432bd0247df66b8e7d066d2" Nov 23 07:10:12 crc kubenswrapper[4559]: I1123 07:10:12.889498 4559 scope.go:117] "RemoveContainer" containerID="cb75c9c2a61182d08852cade52ba64d0dd93308a8c9813f2b07c5c968f748421" Nov 23 07:10:12 crc kubenswrapper[4559]: I1123 07:10:12.907581 4559 scope.go:117] "RemoveContainer" containerID="61609ef0c1efb69623deb3e9eb896c040656881ea7d6aae5d9c4908101dd6605" Nov 23 07:10:12 crc kubenswrapper[4559]: I1123 07:10:12.923629 4559 scope.go:117] "RemoveContainer" containerID="bbd5eebeecbdf13f45dd156157392985c3a4a276178b65794a062e0829d3d543" Nov 23 07:10:12 crc kubenswrapper[4559]: I1123 07:10:12.961715 4559 scope.go:117] "RemoveContainer" containerID="4b7bde1d24163aa6ba263b550d46b5c17d85dad13c90a85a22611065fe68b406" Nov 23 07:10:15 crc kubenswrapper[4559]: I1123 07:10:15.274597 4559 scope.go:117] "RemoveContainer" containerID="ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b" Nov 23 07:10:15 crc kubenswrapper[4559]: E1123 07:10:15.275762 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:10:30 crc kubenswrapper[4559]: I1123 07:10:30.273790 4559 scope.go:117] "RemoveContainer" containerID="ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b" Nov 23 07:10:30 crc kubenswrapper[4559]: E1123 07:10:30.274285 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:10:38 crc kubenswrapper[4559]: I1123 07:10:38.064306 4559 generic.go:334] "Generic (PLEG): container finished" podID="f3cbccfd-3b9d-4d8d-aec1-702b2e494667" containerID="07b908b4162cfb8b79b65c179b46ba62fe994ea5d7de6a2e8808a1de13f27098" exitCode=0 Nov 23 07:10:38 crc kubenswrapper[4559]: I1123 07:10:38.064381 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms" event={"ID":"f3cbccfd-3b9d-4d8d-aec1-702b2e494667","Type":"ContainerDied","Data":"07b908b4162cfb8b79b65c179b46ba62fe994ea5d7de6a2e8808a1de13f27098"} Nov 23 07:10:39 crc kubenswrapper[4559]: I1123 07:10:39.030378 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-tp852"] Nov 23 07:10:39 crc kubenswrapper[4559]: I1123 07:10:39.035761 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-974e-account-create-qw2ds"] Nov 23 07:10:39 crc kubenswrapper[4559]: I1123 07:10:39.042347 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-t898j"] Nov 23 07:10:39 crc kubenswrapper[4559]: I1123 07:10:39.047924 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-d2d8-account-create-tfkjf"] Nov 23 07:10:39 crc kubenswrapper[4559]: I1123 07:10:39.053188 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-kktcm"] Nov 23 07:10:39 crc kubenswrapper[4559]: I1123 07:10:39.058167 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-tp852"] Nov 23 07:10:39 crc kubenswrapper[4559]: I1123 07:10:39.062732 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-974e-account-create-qw2ds"] Nov 23 07:10:39 crc kubenswrapper[4559]: I1123 07:10:39.067912 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-ade7-account-create-mj2zg"] Nov 23 07:10:39 crc kubenswrapper[4559]: I1123 07:10:39.085043 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-t898j"] Nov 23 07:10:39 crc kubenswrapper[4559]: I1123 07:10:39.090805 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-kktcm"] Nov 23 07:10:39 crc kubenswrapper[4559]: I1123 07:10:39.097453 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-d2d8-account-create-tfkjf"] Nov 23 07:10:39 crc kubenswrapper[4559]: I1123 07:10:39.103898 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-ade7-account-create-mj2zg"] Nov 23 07:10:39 crc kubenswrapper[4559]: I1123 07:10:39.381065 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms" Nov 23 07:10:39 crc kubenswrapper[4559]: I1123 07:10:39.548449 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ft2db\" (UniqueName: \"kubernetes.io/projected/f3cbccfd-3b9d-4d8d-aec1-702b2e494667-kube-api-access-ft2db\") pod \"f3cbccfd-3b9d-4d8d-aec1-702b2e494667\" (UID: \"f3cbccfd-3b9d-4d8d-aec1-702b2e494667\") " Nov 23 07:10:39 crc kubenswrapper[4559]: I1123 07:10:39.548481 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3cbccfd-3b9d-4d8d-aec1-702b2e494667-inventory\") pod \"f3cbccfd-3b9d-4d8d-aec1-702b2e494667\" (UID: \"f3cbccfd-3b9d-4d8d-aec1-702b2e494667\") " Nov 23 07:10:39 crc kubenswrapper[4559]: I1123 07:10:39.548553 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f3cbccfd-3b9d-4d8d-aec1-702b2e494667-ssh-key\") pod \"f3cbccfd-3b9d-4d8d-aec1-702b2e494667\" (UID: \"f3cbccfd-3b9d-4d8d-aec1-702b2e494667\") " Nov 23 07:10:39 crc kubenswrapper[4559]: I1123 07:10:39.552586 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3cbccfd-3b9d-4d8d-aec1-702b2e494667-kube-api-access-ft2db" (OuterVolumeSpecName: "kube-api-access-ft2db") pod "f3cbccfd-3b9d-4d8d-aec1-702b2e494667" (UID: "f3cbccfd-3b9d-4d8d-aec1-702b2e494667"). InnerVolumeSpecName "kube-api-access-ft2db". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:10:39 crc kubenswrapper[4559]: I1123 07:10:39.568066 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3cbccfd-3b9d-4d8d-aec1-702b2e494667-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f3cbccfd-3b9d-4d8d-aec1-702b2e494667" (UID: "f3cbccfd-3b9d-4d8d-aec1-702b2e494667"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:39 crc kubenswrapper[4559]: I1123 07:10:39.569290 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3cbccfd-3b9d-4d8d-aec1-702b2e494667-inventory" (OuterVolumeSpecName: "inventory") pod "f3cbccfd-3b9d-4d8d-aec1-702b2e494667" (UID: "f3cbccfd-3b9d-4d8d-aec1-702b2e494667"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:39 crc kubenswrapper[4559]: I1123 07:10:39.650455 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ft2db\" (UniqueName: \"kubernetes.io/projected/f3cbccfd-3b9d-4d8d-aec1-702b2e494667-kube-api-access-ft2db\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:39 crc kubenswrapper[4559]: I1123 07:10:39.650477 4559 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3cbccfd-3b9d-4d8d-aec1-702b2e494667-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:39 crc kubenswrapper[4559]: I1123 07:10:39.650486 4559 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f3cbccfd-3b9d-4d8d-aec1-702b2e494667-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.082500 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms" event={"ID":"f3cbccfd-3b9d-4d8d-aec1-702b2e494667","Type":"ContainerDied","Data":"83562b3814ebe0fc8529074af0d1ad41adca45cc34017cc6b6eaa16538d542b1"} Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.082532 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83562b3814ebe0fc8529074af0d1ad41adca45cc34017cc6b6eaa16538d542b1" Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.082534 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms" Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.133037 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-sw6gw"] Nov 23 07:10:40 crc kubenswrapper[4559]: E1123 07:10:40.133439 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3cbccfd-3b9d-4d8d-aec1-702b2e494667" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.133458 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3cbccfd-3b9d-4d8d-aec1-702b2e494667" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.133680 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3cbccfd-3b9d-4d8d-aec1-702b2e494667" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.134309 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-sw6gw" Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.138219 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.138381 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.139502 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.139906 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mj6c" Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.146436 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-sw6gw"] Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.157704 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcn6x\" (UniqueName: \"kubernetes.io/projected/3f1151cd-5698-4fd5-ba25-e28b48717cb8-kube-api-access-kcn6x\") pod \"ssh-known-hosts-edpm-deployment-sw6gw\" (UID: \"3f1151cd-5698-4fd5-ba25-e28b48717cb8\") " pod="openstack/ssh-known-hosts-edpm-deployment-sw6gw" Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.157763 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3f1151cd-5698-4fd5-ba25-e28b48717cb8-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-sw6gw\" (UID: \"3f1151cd-5698-4fd5-ba25-e28b48717cb8\") " pod="openstack/ssh-known-hosts-edpm-deployment-sw6gw" Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.157795 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3f1151cd-5698-4fd5-ba25-e28b48717cb8-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-sw6gw\" (UID: \"3f1151cd-5698-4fd5-ba25-e28b48717cb8\") " pod="openstack/ssh-known-hosts-edpm-deployment-sw6gw" Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.259421 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcn6x\" (UniqueName: \"kubernetes.io/projected/3f1151cd-5698-4fd5-ba25-e28b48717cb8-kube-api-access-kcn6x\") pod \"ssh-known-hosts-edpm-deployment-sw6gw\" (UID: \"3f1151cd-5698-4fd5-ba25-e28b48717cb8\") " pod="openstack/ssh-known-hosts-edpm-deployment-sw6gw" Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.259496 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3f1151cd-5698-4fd5-ba25-e28b48717cb8-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-sw6gw\" (UID: \"3f1151cd-5698-4fd5-ba25-e28b48717cb8\") " pod="openstack/ssh-known-hosts-edpm-deployment-sw6gw" Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.259532 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3f1151cd-5698-4fd5-ba25-e28b48717cb8-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-sw6gw\" (UID: \"3f1151cd-5698-4fd5-ba25-e28b48717cb8\") " pod="openstack/ssh-known-hosts-edpm-deployment-sw6gw" Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.262458 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3f1151cd-5698-4fd5-ba25-e28b48717cb8-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-sw6gw\" (UID: \"3f1151cd-5698-4fd5-ba25-e28b48717cb8\") " pod="openstack/ssh-known-hosts-edpm-deployment-sw6gw" Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.262465 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3f1151cd-5698-4fd5-ba25-e28b48717cb8-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-sw6gw\" (UID: \"3f1151cd-5698-4fd5-ba25-e28b48717cb8\") " pod="openstack/ssh-known-hosts-edpm-deployment-sw6gw" Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.272716 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcn6x\" (UniqueName: \"kubernetes.io/projected/3f1151cd-5698-4fd5-ba25-e28b48717cb8-kube-api-access-kcn6x\") pod \"ssh-known-hosts-edpm-deployment-sw6gw\" (UID: \"3f1151cd-5698-4fd5-ba25-e28b48717cb8\") " pod="openstack/ssh-known-hosts-edpm-deployment-sw6gw" Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.281735 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="096bd741-6394-4f58-9024-d425d96a0bd7" path="/var/lib/kubelet/pods/096bd741-6394-4f58-9024-d425d96a0bd7/volumes" Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.282271 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78fcad7c-4049-4dae-9fa1-2f8353447e12" path="/var/lib/kubelet/pods/78fcad7c-4049-4dae-9fa1-2f8353447e12/volumes" Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.282793 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79afaafd-5f51-4862-9c09-e7f817d54f73" path="/var/lib/kubelet/pods/79afaafd-5f51-4862-9c09-e7f817d54f73/volumes" Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.283282 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1d3e297-8696-43ae-84f8-da003e5a03e5" path="/var/lib/kubelet/pods/b1d3e297-8696-43ae-84f8-da003e5a03e5/volumes" Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.284204 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec53c68f-1637-47e8-8941-0ada991639ec" path="/var/lib/kubelet/pods/ec53c68f-1637-47e8-8941-0ada991639ec/volumes" Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.284695 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4" path="/var/lib/kubelet/pods/fe7a38e1-2fbe-4cdd-83d1-8bba085a21e4/volumes" Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.447664 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-sw6gw" Nov 23 07:10:40 crc kubenswrapper[4559]: I1123 07:10:40.865801 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-sw6gw"] Nov 23 07:10:41 crc kubenswrapper[4559]: I1123 07:10:41.089254 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-sw6gw" event={"ID":"3f1151cd-5698-4fd5-ba25-e28b48717cb8","Type":"ContainerStarted","Data":"c2f81057e2b1f258755bb256655175ff26843d51a5cc12026d2d2e9e93c9f4d6"} Nov 23 07:10:41 crc kubenswrapper[4559]: I1123 07:10:41.273211 4559 scope.go:117] "RemoveContainer" containerID="ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b" Nov 23 07:10:41 crc kubenswrapper[4559]: E1123 07:10:41.273514 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:10:42 crc kubenswrapper[4559]: I1123 07:10:42.097418 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-sw6gw" event={"ID":"3f1151cd-5698-4fd5-ba25-e28b48717cb8","Type":"ContainerStarted","Data":"eb20d92c33fe4573ebe32da14963871f0707159916eade8dbc7dc478726514fa"} Nov 23 07:10:42 crc kubenswrapper[4559]: I1123 07:10:42.110600 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-sw6gw" podStartSLOduration=1.61005523 podStartE2EDuration="2.110587898s" podCreationTimestamp="2025-11-23 07:10:40 +0000 UTC" firstStartedPulling="2025-11-23 07:10:40.869633606 +0000 UTC m=+1542.891619219" lastFinishedPulling="2025-11-23 07:10:41.370166273 +0000 UTC m=+1543.392151887" observedRunningTime="2025-11-23 07:10:42.108503609 +0000 UTC m=+1544.130489223" watchObservedRunningTime="2025-11-23 07:10:42.110587898 +0000 UTC m=+1544.132573512" Nov 23 07:10:47 crc kubenswrapper[4559]: I1123 07:10:47.127982 4559 generic.go:334] "Generic (PLEG): container finished" podID="3f1151cd-5698-4fd5-ba25-e28b48717cb8" containerID="eb20d92c33fe4573ebe32da14963871f0707159916eade8dbc7dc478726514fa" exitCode=0 Nov 23 07:10:47 crc kubenswrapper[4559]: I1123 07:10:47.128067 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-sw6gw" event={"ID":"3f1151cd-5698-4fd5-ba25-e28b48717cb8","Type":"ContainerDied","Data":"eb20d92c33fe4573ebe32da14963871f0707159916eade8dbc7dc478726514fa"} Nov 23 07:10:48 crc kubenswrapper[4559]: I1123 07:10:48.436675 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-sw6gw" Nov 23 07:10:48 crc kubenswrapper[4559]: I1123 07:10:48.579738 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kcn6x\" (UniqueName: \"kubernetes.io/projected/3f1151cd-5698-4fd5-ba25-e28b48717cb8-kube-api-access-kcn6x\") pod \"3f1151cd-5698-4fd5-ba25-e28b48717cb8\" (UID: \"3f1151cd-5698-4fd5-ba25-e28b48717cb8\") " Nov 23 07:10:48 crc kubenswrapper[4559]: I1123 07:10:48.579786 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3f1151cd-5698-4fd5-ba25-e28b48717cb8-inventory-0\") pod \"3f1151cd-5698-4fd5-ba25-e28b48717cb8\" (UID: \"3f1151cd-5698-4fd5-ba25-e28b48717cb8\") " Nov 23 07:10:48 crc kubenswrapper[4559]: I1123 07:10:48.579826 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3f1151cd-5698-4fd5-ba25-e28b48717cb8-ssh-key-openstack-edpm-ipam\") pod \"3f1151cd-5698-4fd5-ba25-e28b48717cb8\" (UID: \"3f1151cd-5698-4fd5-ba25-e28b48717cb8\") " Nov 23 07:10:48 crc kubenswrapper[4559]: I1123 07:10:48.592293 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f1151cd-5698-4fd5-ba25-e28b48717cb8-kube-api-access-kcn6x" (OuterVolumeSpecName: "kube-api-access-kcn6x") pod "3f1151cd-5698-4fd5-ba25-e28b48717cb8" (UID: "3f1151cd-5698-4fd5-ba25-e28b48717cb8"). InnerVolumeSpecName "kube-api-access-kcn6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:10:48 crc kubenswrapper[4559]: I1123 07:10:48.609847 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f1151cd-5698-4fd5-ba25-e28b48717cb8-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "3f1151cd-5698-4fd5-ba25-e28b48717cb8" (UID: "3f1151cd-5698-4fd5-ba25-e28b48717cb8"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:48 crc kubenswrapper[4559]: I1123 07:10:48.615813 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f1151cd-5698-4fd5-ba25-e28b48717cb8-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "3f1151cd-5698-4fd5-ba25-e28b48717cb8" (UID: "3f1151cd-5698-4fd5-ba25-e28b48717cb8"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:48 crc kubenswrapper[4559]: I1123 07:10:48.681786 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kcn6x\" (UniqueName: \"kubernetes.io/projected/3f1151cd-5698-4fd5-ba25-e28b48717cb8-kube-api-access-kcn6x\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:48 crc kubenswrapper[4559]: I1123 07:10:48.682020 4559 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3f1151cd-5698-4fd5-ba25-e28b48717cb8-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:48 crc kubenswrapper[4559]: I1123 07:10:48.682030 4559 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3f1151cd-5698-4fd5-ba25-e28b48717cb8-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:49 crc kubenswrapper[4559]: I1123 07:10:49.142759 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-sw6gw" event={"ID":"3f1151cd-5698-4fd5-ba25-e28b48717cb8","Type":"ContainerDied","Data":"c2f81057e2b1f258755bb256655175ff26843d51a5cc12026d2d2e9e93c9f4d6"} Nov 23 07:10:49 crc kubenswrapper[4559]: I1123 07:10:49.142797 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2f81057e2b1f258755bb256655175ff26843d51a5cc12026d2d2e9e93c9f4d6" Nov 23 07:10:49 crc kubenswrapper[4559]: I1123 07:10:49.142809 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-sw6gw" Nov 23 07:10:49 crc kubenswrapper[4559]: I1123 07:10:49.196196 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-5flwd"] Nov 23 07:10:49 crc kubenswrapper[4559]: E1123 07:10:49.196744 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f1151cd-5698-4fd5-ba25-e28b48717cb8" containerName="ssh-known-hosts-edpm-deployment" Nov 23 07:10:49 crc kubenswrapper[4559]: I1123 07:10:49.196825 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f1151cd-5698-4fd5-ba25-e28b48717cb8" containerName="ssh-known-hosts-edpm-deployment" Nov 23 07:10:49 crc kubenswrapper[4559]: I1123 07:10:49.197160 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f1151cd-5698-4fd5-ba25-e28b48717cb8" containerName="ssh-known-hosts-edpm-deployment" Nov 23 07:10:49 crc kubenswrapper[4559]: I1123 07:10:49.197793 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5flwd" Nov 23 07:10:49 crc kubenswrapper[4559]: I1123 07:10:49.200473 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 07:10:49 crc kubenswrapper[4559]: I1123 07:10:49.200496 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 07:10:49 crc kubenswrapper[4559]: I1123 07:10:49.200830 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mj6c" Nov 23 07:10:49 crc kubenswrapper[4559]: I1123 07:10:49.200848 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 07:10:49 crc kubenswrapper[4559]: I1123 07:10:49.206417 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-5flwd"] Nov 23 07:10:49 crc kubenswrapper[4559]: I1123 07:10:49.393258 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f91e51c1-c500-4da2-acba-fcdfa134d397-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-5flwd\" (UID: \"f91e51c1-c500-4da2-acba-fcdfa134d397\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5flwd" Nov 23 07:10:49 crc kubenswrapper[4559]: I1123 07:10:49.393980 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f91e51c1-c500-4da2-acba-fcdfa134d397-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-5flwd\" (UID: \"f91e51c1-c500-4da2-acba-fcdfa134d397\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5flwd" Nov 23 07:10:49 crc kubenswrapper[4559]: I1123 07:10:49.394060 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwrqx\" (UniqueName: \"kubernetes.io/projected/f91e51c1-c500-4da2-acba-fcdfa134d397-kube-api-access-nwrqx\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-5flwd\" (UID: \"f91e51c1-c500-4da2-acba-fcdfa134d397\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5flwd" Nov 23 07:10:49 crc kubenswrapper[4559]: I1123 07:10:49.495713 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwrqx\" (UniqueName: \"kubernetes.io/projected/f91e51c1-c500-4da2-acba-fcdfa134d397-kube-api-access-nwrqx\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-5flwd\" (UID: \"f91e51c1-c500-4da2-acba-fcdfa134d397\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5flwd" Nov 23 07:10:49 crc kubenswrapper[4559]: I1123 07:10:49.495831 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f91e51c1-c500-4da2-acba-fcdfa134d397-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-5flwd\" (UID: \"f91e51c1-c500-4da2-acba-fcdfa134d397\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5flwd" Nov 23 07:10:49 crc kubenswrapper[4559]: I1123 07:10:49.495933 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f91e51c1-c500-4da2-acba-fcdfa134d397-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-5flwd\" (UID: \"f91e51c1-c500-4da2-acba-fcdfa134d397\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5flwd" Nov 23 07:10:49 crc kubenswrapper[4559]: I1123 07:10:49.499549 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f91e51c1-c500-4da2-acba-fcdfa134d397-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-5flwd\" (UID: \"f91e51c1-c500-4da2-acba-fcdfa134d397\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5flwd" Nov 23 07:10:49 crc kubenswrapper[4559]: I1123 07:10:49.499791 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f91e51c1-c500-4da2-acba-fcdfa134d397-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-5flwd\" (UID: \"f91e51c1-c500-4da2-acba-fcdfa134d397\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5flwd" Nov 23 07:10:49 crc kubenswrapper[4559]: I1123 07:10:49.509948 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwrqx\" (UniqueName: \"kubernetes.io/projected/f91e51c1-c500-4da2-acba-fcdfa134d397-kube-api-access-nwrqx\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-5flwd\" (UID: \"f91e51c1-c500-4da2-acba-fcdfa134d397\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5flwd" Nov 23 07:10:49 crc kubenswrapper[4559]: I1123 07:10:49.511626 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5flwd" Nov 23 07:10:49 crc kubenswrapper[4559]: I1123 07:10:49.930194 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-5flwd"] Nov 23 07:10:50 crc kubenswrapper[4559]: I1123 07:10:50.151865 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5flwd" event={"ID":"f91e51c1-c500-4da2-acba-fcdfa134d397","Type":"ContainerStarted","Data":"cd33b4e3a22db17714cdf8d59e675b56b01016508de62a76fda74f3ed0b00ab9"} Nov 23 07:10:51 crc kubenswrapper[4559]: I1123 07:10:51.158908 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5flwd" event={"ID":"f91e51c1-c500-4da2-acba-fcdfa134d397","Type":"ContainerStarted","Data":"d139c1b2e9b4875eae819b1ad7fcad1c8aef0c39ae9a64f1568987fbb643ee2c"} Nov 23 07:10:51 crc kubenswrapper[4559]: I1123 07:10:51.179045 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5flwd" podStartSLOduration=1.603935407 podStartE2EDuration="2.179033069s" podCreationTimestamp="2025-11-23 07:10:49 +0000 UTC" firstStartedPulling="2025-11-23 07:10:49.935675797 +0000 UTC m=+1551.957661411" lastFinishedPulling="2025-11-23 07:10:50.51077346 +0000 UTC m=+1552.532759073" observedRunningTime="2025-11-23 07:10:51.169975085 +0000 UTC m=+1553.191960700" watchObservedRunningTime="2025-11-23 07:10:51.179033069 +0000 UTC m=+1553.201018683" Nov 23 07:10:53 crc kubenswrapper[4559]: I1123 07:10:53.273554 4559 scope.go:117] "RemoveContainer" containerID="ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b" Nov 23 07:10:53 crc kubenswrapper[4559]: E1123 07:10:53.274087 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:10:56 crc kubenswrapper[4559]: I1123 07:10:56.188690 4559 generic.go:334] "Generic (PLEG): container finished" podID="f91e51c1-c500-4da2-acba-fcdfa134d397" containerID="d139c1b2e9b4875eae819b1ad7fcad1c8aef0c39ae9a64f1568987fbb643ee2c" exitCode=0 Nov 23 07:10:56 crc kubenswrapper[4559]: I1123 07:10:56.188765 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5flwd" event={"ID":"f91e51c1-c500-4da2-acba-fcdfa134d397","Type":"ContainerDied","Data":"d139c1b2e9b4875eae819b1ad7fcad1c8aef0c39ae9a64f1568987fbb643ee2c"} Nov 23 07:10:57 crc kubenswrapper[4559]: I1123 07:10:57.497913 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5flwd" Nov 23 07:10:57 crc kubenswrapper[4559]: I1123 07:10:57.630392 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f91e51c1-c500-4da2-acba-fcdfa134d397-inventory\") pod \"f91e51c1-c500-4da2-acba-fcdfa134d397\" (UID: \"f91e51c1-c500-4da2-acba-fcdfa134d397\") " Nov 23 07:10:57 crc kubenswrapper[4559]: I1123 07:10:57.630656 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwrqx\" (UniqueName: \"kubernetes.io/projected/f91e51c1-c500-4da2-acba-fcdfa134d397-kube-api-access-nwrqx\") pod \"f91e51c1-c500-4da2-acba-fcdfa134d397\" (UID: \"f91e51c1-c500-4da2-acba-fcdfa134d397\") " Nov 23 07:10:57 crc kubenswrapper[4559]: I1123 07:10:57.630692 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f91e51c1-c500-4da2-acba-fcdfa134d397-ssh-key\") pod \"f91e51c1-c500-4da2-acba-fcdfa134d397\" (UID: \"f91e51c1-c500-4da2-acba-fcdfa134d397\") " Nov 23 07:10:57 crc kubenswrapper[4559]: I1123 07:10:57.647109 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f91e51c1-c500-4da2-acba-fcdfa134d397-kube-api-access-nwrqx" (OuterVolumeSpecName: "kube-api-access-nwrqx") pod "f91e51c1-c500-4da2-acba-fcdfa134d397" (UID: "f91e51c1-c500-4da2-acba-fcdfa134d397"). InnerVolumeSpecName "kube-api-access-nwrqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:10:57 crc kubenswrapper[4559]: I1123 07:10:57.650622 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f91e51c1-c500-4da2-acba-fcdfa134d397-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f91e51c1-c500-4da2-acba-fcdfa134d397" (UID: "f91e51c1-c500-4da2-acba-fcdfa134d397"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:57 crc kubenswrapper[4559]: I1123 07:10:57.650942 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f91e51c1-c500-4da2-acba-fcdfa134d397-inventory" (OuterVolumeSpecName: "inventory") pod "f91e51c1-c500-4da2-acba-fcdfa134d397" (UID: "f91e51c1-c500-4da2-acba-fcdfa134d397"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:10:57 crc kubenswrapper[4559]: I1123 07:10:57.732827 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwrqx\" (UniqueName: \"kubernetes.io/projected/f91e51c1-c500-4da2-acba-fcdfa134d397-kube-api-access-nwrqx\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:57 crc kubenswrapper[4559]: I1123 07:10:57.732858 4559 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f91e51c1-c500-4da2-acba-fcdfa134d397-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:57 crc kubenswrapper[4559]: I1123 07:10:57.732869 4559 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f91e51c1-c500-4da2-acba-fcdfa134d397-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 07:10:58 crc kubenswrapper[4559]: I1123 07:10:58.201832 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5flwd" event={"ID":"f91e51c1-c500-4da2-acba-fcdfa134d397","Type":"ContainerDied","Data":"cd33b4e3a22db17714cdf8d59e675b56b01016508de62a76fda74f3ed0b00ab9"} Nov 23 07:10:58 crc kubenswrapper[4559]: I1123 07:10:58.202024 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd33b4e3a22db17714cdf8d59e675b56b01016508de62a76fda74f3ed0b00ab9" Nov 23 07:10:58 crc kubenswrapper[4559]: I1123 07:10:58.201858 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5flwd" Nov 23 07:10:58 crc kubenswrapper[4559]: I1123 07:10:58.248835 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2"] Nov 23 07:10:58 crc kubenswrapper[4559]: E1123 07:10:58.249271 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f91e51c1-c500-4da2-acba-fcdfa134d397" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 23 07:10:58 crc kubenswrapper[4559]: I1123 07:10:58.249334 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="f91e51c1-c500-4da2-acba-fcdfa134d397" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 23 07:10:58 crc kubenswrapper[4559]: I1123 07:10:58.249558 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="f91e51c1-c500-4da2-acba-fcdfa134d397" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 23 07:10:58 crc kubenswrapper[4559]: I1123 07:10:58.250180 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2" Nov 23 07:10:58 crc kubenswrapper[4559]: I1123 07:10:58.251986 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 07:10:58 crc kubenswrapper[4559]: I1123 07:10:58.252062 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 07:10:58 crc kubenswrapper[4559]: I1123 07:10:58.252394 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mj6c" Nov 23 07:10:58 crc kubenswrapper[4559]: I1123 07:10:58.255460 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 07:10:58 crc kubenswrapper[4559]: I1123 07:10:58.257329 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2"] Nov 23 07:10:58 crc kubenswrapper[4559]: I1123 07:10:58.341393 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdc8h\" (UniqueName: \"kubernetes.io/projected/67ea0e44-543c-4601-88e5-51ee0d43424a-kube-api-access-kdc8h\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2\" (UID: \"67ea0e44-543c-4601-88e5-51ee0d43424a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2" Nov 23 07:10:58 crc kubenswrapper[4559]: I1123 07:10:58.341445 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67ea0e44-543c-4601-88e5-51ee0d43424a-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2\" (UID: \"67ea0e44-543c-4601-88e5-51ee0d43424a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2" Nov 23 07:10:58 crc kubenswrapper[4559]: I1123 07:10:58.341471 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67ea0e44-543c-4601-88e5-51ee0d43424a-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2\" (UID: \"67ea0e44-543c-4601-88e5-51ee0d43424a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2" Nov 23 07:10:58 crc kubenswrapper[4559]: I1123 07:10:58.442505 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdc8h\" (UniqueName: \"kubernetes.io/projected/67ea0e44-543c-4601-88e5-51ee0d43424a-kube-api-access-kdc8h\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2\" (UID: \"67ea0e44-543c-4601-88e5-51ee0d43424a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2" Nov 23 07:10:58 crc kubenswrapper[4559]: I1123 07:10:58.442558 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67ea0e44-543c-4601-88e5-51ee0d43424a-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2\" (UID: \"67ea0e44-543c-4601-88e5-51ee0d43424a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2" Nov 23 07:10:58 crc kubenswrapper[4559]: I1123 07:10:58.442592 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67ea0e44-543c-4601-88e5-51ee0d43424a-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2\" (UID: \"67ea0e44-543c-4601-88e5-51ee0d43424a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2" Nov 23 07:10:58 crc kubenswrapper[4559]: I1123 07:10:58.455093 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67ea0e44-543c-4601-88e5-51ee0d43424a-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2\" (UID: \"67ea0e44-543c-4601-88e5-51ee0d43424a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2" Nov 23 07:10:58 crc kubenswrapper[4559]: I1123 07:10:58.455104 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67ea0e44-543c-4601-88e5-51ee0d43424a-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2\" (UID: \"67ea0e44-543c-4601-88e5-51ee0d43424a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2" Nov 23 07:10:58 crc kubenswrapper[4559]: I1123 07:10:58.457405 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdc8h\" (UniqueName: \"kubernetes.io/projected/67ea0e44-543c-4601-88e5-51ee0d43424a-kube-api-access-kdc8h\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2\" (UID: \"67ea0e44-543c-4601-88e5-51ee0d43424a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2" Nov 23 07:10:58 crc kubenswrapper[4559]: I1123 07:10:58.564307 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2" Nov 23 07:10:58 crc kubenswrapper[4559]: I1123 07:10:58.989561 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2"] Nov 23 07:10:59 crc kubenswrapper[4559]: I1123 07:10:59.208923 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2" event={"ID":"67ea0e44-543c-4601-88e5-51ee0d43424a","Type":"ContainerStarted","Data":"0e0b23073df02599d3cc406e7b5ed6df4ad51aa87865e7ca86562498bc7c7192"} Nov 23 07:11:00 crc kubenswrapper[4559]: I1123 07:11:00.029017 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7nhf7"] Nov 23 07:11:00 crc kubenswrapper[4559]: I1123 07:11:00.035355 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7nhf7"] Nov 23 07:11:00 crc kubenswrapper[4559]: I1123 07:11:00.215579 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2" event={"ID":"67ea0e44-543c-4601-88e5-51ee0d43424a","Type":"ContainerStarted","Data":"228166ad692d482b00106f248cba871317a6b32c0f589b273d832c1d9e6e8382"} Nov 23 07:11:00 crc kubenswrapper[4559]: I1123 07:11:00.231054 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2" podStartSLOduration=1.755952089 podStartE2EDuration="2.231041259s" podCreationTimestamp="2025-11-23 07:10:58 +0000 UTC" firstStartedPulling="2025-11-23 07:10:58.995400047 +0000 UTC m=+1561.017385661" lastFinishedPulling="2025-11-23 07:10:59.470489217 +0000 UTC m=+1561.492474831" observedRunningTime="2025-11-23 07:11:00.227522923 +0000 UTC m=+1562.249508537" watchObservedRunningTime="2025-11-23 07:11:00.231041259 +0000 UTC m=+1562.253026873" Nov 23 07:11:00 crc kubenswrapper[4559]: I1123 07:11:00.282251 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b08ced2f-75ec-4b10-9e87-e83572224054" path="/var/lib/kubelet/pods/b08ced2f-75ec-4b10-9e87-e83572224054/volumes" Nov 23 07:11:06 crc kubenswrapper[4559]: I1123 07:11:06.274073 4559 scope.go:117] "RemoveContainer" containerID="ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b" Nov 23 07:11:06 crc kubenswrapper[4559]: E1123 07:11:06.274715 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:11:07 crc kubenswrapper[4559]: I1123 07:11:07.261012 4559 generic.go:334] "Generic (PLEG): container finished" podID="67ea0e44-543c-4601-88e5-51ee0d43424a" containerID="228166ad692d482b00106f248cba871317a6b32c0f589b273d832c1d9e6e8382" exitCode=0 Nov 23 07:11:07 crc kubenswrapper[4559]: I1123 07:11:07.261103 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2" event={"ID":"67ea0e44-543c-4601-88e5-51ee0d43424a","Type":"ContainerDied","Data":"228166ad692d482b00106f248cba871317a6b32c0f589b273d832c1d9e6e8382"} Nov 23 07:11:08 crc kubenswrapper[4559]: I1123 07:11:08.544561 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2" Nov 23 07:11:08 crc kubenswrapper[4559]: I1123 07:11:08.693103 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdc8h\" (UniqueName: \"kubernetes.io/projected/67ea0e44-543c-4601-88e5-51ee0d43424a-kube-api-access-kdc8h\") pod \"67ea0e44-543c-4601-88e5-51ee0d43424a\" (UID: \"67ea0e44-543c-4601-88e5-51ee0d43424a\") " Nov 23 07:11:08 crc kubenswrapper[4559]: I1123 07:11:08.693212 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67ea0e44-543c-4601-88e5-51ee0d43424a-ssh-key\") pod \"67ea0e44-543c-4601-88e5-51ee0d43424a\" (UID: \"67ea0e44-543c-4601-88e5-51ee0d43424a\") " Nov 23 07:11:08 crc kubenswrapper[4559]: I1123 07:11:08.693381 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67ea0e44-543c-4601-88e5-51ee0d43424a-inventory\") pod \"67ea0e44-543c-4601-88e5-51ee0d43424a\" (UID: \"67ea0e44-543c-4601-88e5-51ee0d43424a\") " Nov 23 07:11:08 crc kubenswrapper[4559]: I1123 07:11:08.710133 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67ea0e44-543c-4601-88e5-51ee0d43424a-kube-api-access-kdc8h" (OuterVolumeSpecName: "kube-api-access-kdc8h") pod "67ea0e44-543c-4601-88e5-51ee0d43424a" (UID: "67ea0e44-543c-4601-88e5-51ee0d43424a"). InnerVolumeSpecName "kube-api-access-kdc8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:11:08 crc kubenswrapper[4559]: I1123 07:11:08.714434 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67ea0e44-543c-4601-88e5-51ee0d43424a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "67ea0e44-543c-4601-88e5-51ee0d43424a" (UID: "67ea0e44-543c-4601-88e5-51ee0d43424a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:08 crc kubenswrapper[4559]: I1123 07:11:08.715525 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67ea0e44-543c-4601-88e5-51ee0d43424a-inventory" (OuterVolumeSpecName: "inventory") pod "67ea0e44-543c-4601-88e5-51ee0d43424a" (UID: "67ea0e44-543c-4601-88e5-51ee0d43424a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:08 crc kubenswrapper[4559]: I1123 07:11:08.795335 4559 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67ea0e44-543c-4601-88e5-51ee0d43424a-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:08 crc kubenswrapper[4559]: I1123 07:11:08.795358 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdc8h\" (UniqueName: \"kubernetes.io/projected/67ea0e44-543c-4601-88e5-51ee0d43424a-kube-api-access-kdc8h\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:08 crc kubenswrapper[4559]: I1123 07:11:08.795369 4559 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67ea0e44-543c-4601-88e5-51ee0d43424a-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.274281 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2" event={"ID":"67ea0e44-543c-4601-88e5-51ee0d43424a","Type":"ContainerDied","Data":"0e0b23073df02599d3cc406e7b5ed6df4ad51aa87865e7ca86562498bc7c7192"} Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.274336 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e0b23073df02599d3cc406e7b5ed6df4ad51aa87865e7ca86562498bc7c7192" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.274336 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.323979 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs"] Nov 23 07:11:09 crc kubenswrapper[4559]: E1123 07:11:09.324488 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67ea0e44-543c-4601-88e5-51ee0d43424a" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.324507 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="67ea0e44-543c-4601-88e5-51ee0d43424a" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.324691 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="67ea0e44-543c-4601-88e5-51ee0d43424a" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.325215 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.330157 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.330350 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.330510 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mj6c" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.330158 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.330725 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.330857 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.330961 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.331114 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.341695 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs"] Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.505764 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.505823 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.505860 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.505896 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.505930 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.505945 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.505961 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.506052 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jgth\" (UniqueName: \"kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-kube-api-access-8jgth\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.506101 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.506143 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.506180 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.506273 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.506295 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.506330 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.607366 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.608094 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.608177 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.608278 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.608452 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.608528 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.608597 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.608714 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.608798 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.608874 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.608942 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.609026 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jgth\" (UniqueName: \"kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-kube-api-access-8jgth\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.609094 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.609167 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.611598 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.611719 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.612061 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.612503 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.612518 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.612568 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.612635 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.613484 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.613582 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.613726 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.613991 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.614334 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.615352 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.624907 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jgth\" (UniqueName: \"kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-kube-api-access-8jgth\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:09 crc kubenswrapper[4559]: I1123 07:11:09.639929 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:10 crc kubenswrapper[4559]: I1123 07:11:10.066697 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs"] Nov 23 07:11:10 crc kubenswrapper[4559]: I1123 07:11:10.281457 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" event={"ID":"aa380021-cb24-43b1-bb0d-e3dbbd5f6676","Type":"ContainerStarted","Data":"e19a07451f4c059b953de6b80f97636489e6b82a517178b648ac5f0b5815875a"} Nov 23 07:11:11 crc kubenswrapper[4559]: I1123 07:11:11.289412 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" event={"ID":"aa380021-cb24-43b1-bb0d-e3dbbd5f6676","Type":"ContainerStarted","Data":"4c677e6b8ecae7d81981b8d2abd19df5d777b25cc3f402162dac029ddd3af7d2"} Nov 23 07:11:13 crc kubenswrapper[4559]: I1123 07:11:13.147128 4559 scope.go:117] "RemoveContainer" containerID="8a0ee1424748e596b9fdc04cc74255047de5f668af4056f42c70810a28a7af69" Nov 23 07:11:13 crc kubenswrapper[4559]: I1123 07:11:13.165323 4559 scope.go:117] "RemoveContainer" containerID="e626d8fb33c8d679398bf1d7cdc6807e0533a46669e270a980cda59efc314eb1" Nov 23 07:11:13 crc kubenswrapper[4559]: I1123 07:11:13.198559 4559 scope.go:117] "RemoveContainer" containerID="dd9cc55d2d749115f38ba72ac2c037e554c026fa3b9dfb051ed05730645e4c3b" Nov 23 07:11:13 crc kubenswrapper[4559]: I1123 07:11:13.245694 4559 scope.go:117] "RemoveContainer" containerID="25386b49570cd3a8fc6b9401d205700597ac4704ba860466f5ad40289cf95742" Nov 23 07:11:13 crc kubenswrapper[4559]: I1123 07:11:13.263538 4559 scope.go:117] "RemoveContainer" containerID="c23d564932376ae305fa54fa32a91e2bae72476f83833fbf5561e5956d153e8f" Nov 23 07:11:13 crc kubenswrapper[4559]: I1123 07:11:13.294754 4559 scope.go:117] "RemoveContainer" containerID="5c1643a15abf22c3a043c168790f7e71a2c2d53736e6df77cc37dfea667c43a7" Nov 23 07:11:13 crc kubenswrapper[4559]: I1123 07:11:13.322297 4559 scope.go:117] "RemoveContainer" containerID="6f027cc235e87ebc22c09ef418681e2c53a6dad2c2eec69ac930bbd1af843d58" Nov 23 07:11:19 crc kubenswrapper[4559]: I1123 07:11:19.017782 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" podStartSLOduration=9.56443544 podStartE2EDuration="10.017767554s" podCreationTimestamp="2025-11-23 07:11:09 +0000 UTC" firstStartedPulling="2025-11-23 07:11:10.072818276 +0000 UTC m=+1572.094803890" lastFinishedPulling="2025-11-23 07:11:10.526150389 +0000 UTC m=+1572.548136004" observedRunningTime="2025-11-23 07:11:11.303782042 +0000 UTC m=+1573.325767656" watchObservedRunningTime="2025-11-23 07:11:19.017767554 +0000 UTC m=+1581.039753168" Nov 23 07:11:19 crc kubenswrapper[4559]: I1123 07:11:19.023900 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-g9svq"] Nov 23 07:11:19 crc kubenswrapper[4559]: I1123 07:11:19.028859 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-g9svq"] Nov 23 07:11:20 crc kubenswrapper[4559]: I1123 07:11:20.273426 4559 scope.go:117] "RemoveContainer" containerID="ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b" Nov 23 07:11:20 crc kubenswrapper[4559]: E1123 07:11:20.273671 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:11:20 crc kubenswrapper[4559]: I1123 07:11:20.281820 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf4b8627-5670-4ab6-af0d-94b20fb48524" path="/var/lib/kubelet/pods/cf4b8627-5670-4ab6-af0d-94b20fb48524/volumes" Nov 23 07:11:21 crc kubenswrapper[4559]: I1123 07:11:21.026857 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-vfsc8"] Nov 23 07:11:21 crc kubenswrapper[4559]: I1123 07:11:21.032417 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-vfsc8"] Nov 23 07:11:22 crc kubenswrapper[4559]: I1123 07:11:22.282138 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="baf469bd-b908-4271-8688-ee06c3b64e0b" path="/var/lib/kubelet/pods/baf469bd-b908-4271-8688-ee06c3b64e0b/volumes" Nov 23 07:11:33 crc kubenswrapper[4559]: I1123 07:11:33.273916 4559 scope.go:117] "RemoveContainer" containerID="ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b" Nov 23 07:11:33 crc kubenswrapper[4559]: E1123 07:11:33.274878 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:11:33 crc kubenswrapper[4559]: I1123 07:11:33.370449 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-r7n69"] Nov 23 07:11:33 crc kubenswrapper[4559]: I1123 07:11:33.372093 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r7n69" Nov 23 07:11:33 crc kubenswrapper[4559]: I1123 07:11:33.381106 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r7n69"] Nov 23 07:11:33 crc kubenswrapper[4559]: I1123 07:11:33.384612 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc47e6ec-9e05-422c-b0af-08ec3016cd47-catalog-content\") pod \"redhat-marketplace-r7n69\" (UID: \"cc47e6ec-9e05-422c-b0af-08ec3016cd47\") " pod="openshift-marketplace/redhat-marketplace-r7n69" Nov 23 07:11:33 crc kubenswrapper[4559]: I1123 07:11:33.385311 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc47e6ec-9e05-422c-b0af-08ec3016cd47-utilities\") pod \"redhat-marketplace-r7n69\" (UID: \"cc47e6ec-9e05-422c-b0af-08ec3016cd47\") " pod="openshift-marketplace/redhat-marketplace-r7n69" Nov 23 07:11:33 crc kubenswrapper[4559]: I1123 07:11:33.385407 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjf6p\" (UniqueName: \"kubernetes.io/projected/cc47e6ec-9e05-422c-b0af-08ec3016cd47-kube-api-access-kjf6p\") pod \"redhat-marketplace-r7n69\" (UID: \"cc47e6ec-9e05-422c-b0af-08ec3016cd47\") " pod="openshift-marketplace/redhat-marketplace-r7n69" Nov 23 07:11:33 crc kubenswrapper[4559]: I1123 07:11:33.486918 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc47e6ec-9e05-422c-b0af-08ec3016cd47-catalog-content\") pod \"redhat-marketplace-r7n69\" (UID: \"cc47e6ec-9e05-422c-b0af-08ec3016cd47\") " pod="openshift-marketplace/redhat-marketplace-r7n69" Nov 23 07:11:33 crc kubenswrapper[4559]: I1123 07:11:33.487160 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc47e6ec-9e05-422c-b0af-08ec3016cd47-utilities\") pod \"redhat-marketplace-r7n69\" (UID: \"cc47e6ec-9e05-422c-b0af-08ec3016cd47\") " pod="openshift-marketplace/redhat-marketplace-r7n69" Nov 23 07:11:33 crc kubenswrapper[4559]: I1123 07:11:33.487253 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjf6p\" (UniqueName: \"kubernetes.io/projected/cc47e6ec-9e05-422c-b0af-08ec3016cd47-kube-api-access-kjf6p\") pod \"redhat-marketplace-r7n69\" (UID: \"cc47e6ec-9e05-422c-b0af-08ec3016cd47\") " pod="openshift-marketplace/redhat-marketplace-r7n69" Nov 23 07:11:33 crc kubenswrapper[4559]: I1123 07:11:33.487510 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc47e6ec-9e05-422c-b0af-08ec3016cd47-catalog-content\") pod \"redhat-marketplace-r7n69\" (UID: \"cc47e6ec-9e05-422c-b0af-08ec3016cd47\") " pod="openshift-marketplace/redhat-marketplace-r7n69" Nov 23 07:11:33 crc kubenswrapper[4559]: I1123 07:11:33.487764 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc47e6ec-9e05-422c-b0af-08ec3016cd47-utilities\") pod \"redhat-marketplace-r7n69\" (UID: \"cc47e6ec-9e05-422c-b0af-08ec3016cd47\") " pod="openshift-marketplace/redhat-marketplace-r7n69" Nov 23 07:11:33 crc kubenswrapper[4559]: I1123 07:11:33.506171 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjf6p\" (UniqueName: \"kubernetes.io/projected/cc47e6ec-9e05-422c-b0af-08ec3016cd47-kube-api-access-kjf6p\") pod \"redhat-marketplace-r7n69\" (UID: \"cc47e6ec-9e05-422c-b0af-08ec3016cd47\") " pod="openshift-marketplace/redhat-marketplace-r7n69" Nov 23 07:11:33 crc kubenswrapper[4559]: I1123 07:11:33.689387 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r7n69" Nov 23 07:11:34 crc kubenswrapper[4559]: I1123 07:11:34.083769 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r7n69"] Nov 23 07:11:34 crc kubenswrapper[4559]: I1123 07:11:34.472686 4559 generic.go:334] "Generic (PLEG): container finished" podID="cc47e6ec-9e05-422c-b0af-08ec3016cd47" containerID="0b8cf628688fe803e8f3376371a3e3f0bfb7b877de4f7ab9aa7c97bfb8ce2a0a" exitCode=0 Nov 23 07:11:34 crc kubenswrapper[4559]: I1123 07:11:34.472783 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r7n69" event={"ID":"cc47e6ec-9e05-422c-b0af-08ec3016cd47","Type":"ContainerDied","Data":"0b8cf628688fe803e8f3376371a3e3f0bfb7b877de4f7ab9aa7c97bfb8ce2a0a"} Nov 23 07:11:34 crc kubenswrapper[4559]: I1123 07:11:34.473006 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r7n69" event={"ID":"cc47e6ec-9e05-422c-b0af-08ec3016cd47","Type":"ContainerStarted","Data":"e77df62d71a12b82a7706716d54b917ba4fe672a76b6e06702edae9d0f2fc7ca"} Nov 23 07:11:35 crc kubenswrapper[4559]: I1123 07:11:35.484365 4559 generic.go:334] "Generic (PLEG): container finished" podID="cc47e6ec-9e05-422c-b0af-08ec3016cd47" containerID="94c4e3a5c2a641bc6fdd936ac6b2309513f9582119d67750cc92c8c9b08064c3" exitCode=0 Nov 23 07:11:35 crc kubenswrapper[4559]: I1123 07:11:35.484560 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r7n69" event={"ID":"cc47e6ec-9e05-422c-b0af-08ec3016cd47","Type":"ContainerDied","Data":"94c4e3a5c2a641bc6fdd936ac6b2309513f9582119d67750cc92c8c9b08064c3"} Nov 23 07:11:36 crc kubenswrapper[4559]: I1123 07:11:36.494116 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r7n69" event={"ID":"cc47e6ec-9e05-422c-b0af-08ec3016cd47","Type":"ContainerStarted","Data":"104bf0aac053715be0752abf67831d67ceead7eecacfa280fa05220b748c8a47"} Nov 23 07:11:36 crc kubenswrapper[4559]: I1123 07:11:36.509245 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-r7n69" podStartSLOduration=2.017532428 podStartE2EDuration="3.509232147s" podCreationTimestamp="2025-11-23 07:11:33 +0000 UTC" firstStartedPulling="2025-11-23 07:11:34.475377316 +0000 UTC m=+1596.497362930" lastFinishedPulling="2025-11-23 07:11:35.967077035 +0000 UTC m=+1597.989062649" observedRunningTime="2025-11-23 07:11:36.505854284 +0000 UTC m=+1598.527839898" watchObservedRunningTime="2025-11-23 07:11:36.509232147 +0000 UTC m=+1598.531217761" Nov 23 07:11:37 crc kubenswrapper[4559]: I1123 07:11:37.501677 4559 generic.go:334] "Generic (PLEG): container finished" podID="aa380021-cb24-43b1-bb0d-e3dbbd5f6676" containerID="4c677e6b8ecae7d81981b8d2abd19df5d777b25cc3f402162dac029ddd3af7d2" exitCode=0 Nov 23 07:11:37 crc kubenswrapper[4559]: I1123 07:11:37.501744 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" event={"ID":"aa380021-cb24-43b1-bb0d-e3dbbd5f6676","Type":"ContainerDied","Data":"4c677e6b8ecae7d81981b8d2abd19df5d777b25cc3f402162dac029ddd3af7d2"} Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.819658 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.868067 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.868104 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-repo-setup-combined-ca-bundle\") pod \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.868136 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-inventory\") pod \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.868155 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.868173 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-ovn-combined-ca-bundle\") pod \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.868193 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-nova-combined-ca-bundle\") pod \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.868228 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-telemetry-combined-ca-bundle\") pod \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.872403 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "aa380021-cb24-43b1-bb0d-e3dbbd5f6676" (UID: "aa380021-cb24-43b1-bb0d-e3dbbd5f6676"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.872414 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "aa380021-cb24-43b1-bb0d-e3dbbd5f6676" (UID: "aa380021-cb24-43b1-bb0d-e3dbbd5f6676"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.872963 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "aa380021-cb24-43b1-bb0d-e3dbbd5f6676" (UID: "aa380021-cb24-43b1-bb0d-e3dbbd5f6676"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.873789 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "aa380021-cb24-43b1-bb0d-e3dbbd5f6676" (UID: "aa380021-cb24-43b1-bb0d-e3dbbd5f6676"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.873895 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "aa380021-cb24-43b1-bb0d-e3dbbd5f6676" (UID: "aa380021-cb24-43b1-bb0d-e3dbbd5f6676"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.874154 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "aa380021-cb24-43b1-bb0d-e3dbbd5f6676" (UID: "aa380021-cb24-43b1-bb0d-e3dbbd5f6676"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.890383 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-inventory" (OuterVolumeSpecName: "inventory") pod "aa380021-cb24-43b1-bb0d-e3dbbd5f6676" (UID: "aa380021-cb24-43b1-bb0d-e3dbbd5f6676"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.969153 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jgth\" (UniqueName: \"kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-kube-api-access-8jgth\") pod \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.969212 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-neutron-metadata-combined-ca-bundle\") pod \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.969371 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-bootstrap-combined-ca-bundle\") pod \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.969424 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-openstack-edpm-ipam-ovn-default-certs-0\") pod \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.969442 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-ssh-key\") pod \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.969457 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-libvirt-combined-ca-bundle\") pod \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.969494 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\" (UID: \"aa380021-cb24-43b1-bb0d-e3dbbd5f6676\") " Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.969745 4559 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.969761 4559 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.969771 4559 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.969781 4559 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.969792 4559 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.969802 4559 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.969810 4559 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.972044 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "aa380021-cb24-43b1-bb0d-e3dbbd5f6676" (UID: "aa380021-cb24-43b1-bb0d-e3dbbd5f6676"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.972997 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-kube-api-access-8jgth" (OuterVolumeSpecName: "kube-api-access-8jgth") pod "aa380021-cb24-43b1-bb0d-e3dbbd5f6676" (UID: "aa380021-cb24-43b1-bb0d-e3dbbd5f6676"). InnerVolumeSpecName "kube-api-access-8jgth". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.973325 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "aa380021-cb24-43b1-bb0d-e3dbbd5f6676" (UID: "aa380021-cb24-43b1-bb0d-e3dbbd5f6676"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.973353 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "aa380021-cb24-43b1-bb0d-e3dbbd5f6676" (UID: "aa380021-cb24-43b1-bb0d-e3dbbd5f6676"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.973386 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "aa380021-cb24-43b1-bb0d-e3dbbd5f6676" (UID: "aa380021-cb24-43b1-bb0d-e3dbbd5f6676"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.973954 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "aa380021-cb24-43b1-bb0d-e3dbbd5f6676" (UID: "aa380021-cb24-43b1-bb0d-e3dbbd5f6676"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:38 crc kubenswrapper[4559]: I1123 07:11:38.988322 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "aa380021-cb24-43b1-bb0d-e3dbbd5f6676" (UID: "aa380021-cb24-43b1-bb0d-e3dbbd5f6676"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.071029 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jgth\" (UniqueName: \"kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-kube-api-access-8jgth\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.071052 4559 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.071062 4559 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.071073 4559 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.071081 4559 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.071088 4559 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.071096 4559 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/aa380021-cb24-43b1-bb0d-e3dbbd5f6676-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.516669 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" event={"ID":"aa380021-cb24-43b1-bb0d-e3dbbd5f6676","Type":"ContainerDied","Data":"e19a07451f4c059b953de6b80f97636489e6b82a517178b648ac5f0b5815875a"} Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.516691 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.516703 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e19a07451f4c059b953de6b80f97636489e6b82a517178b648ac5f0b5815875a" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.610218 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-9svlh"] Nov 23 07:11:39 crc kubenswrapper[4559]: E1123 07:11:39.610571 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa380021-cb24-43b1-bb0d-e3dbbd5f6676" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.610588 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa380021-cb24-43b1-bb0d-e3dbbd5f6676" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.610803 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa380021-cb24-43b1-bb0d-e3dbbd5f6676" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.611334 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9svlh" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.616174 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.616282 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mj6c" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.616356 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.616356 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.623041 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.629148 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-9svlh"] Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.784066 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9svlh\" (UID: \"4f4ac16e-c282-4fa3-bdbe-e5f848782f47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9svlh" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.784106 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9svlh\" (UID: \"4f4ac16e-c282-4fa3-bdbe-e5f848782f47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9svlh" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.784147 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9svlh\" (UID: \"4f4ac16e-c282-4fa3-bdbe-e5f848782f47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9svlh" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.784252 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5whjv\" (UniqueName: \"kubernetes.io/projected/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-kube-api-access-5whjv\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9svlh\" (UID: \"4f4ac16e-c282-4fa3-bdbe-e5f848782f47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9svlh" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.784329 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9svlh\" (UID: \"4f4ac16e-c282-4fa3-bdbe-e5f848782f47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9svlh" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.886795 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9svlh\" (UID: \"4f4ac16e-c282-4fa3-bdbe-e5f848782f47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9svlh" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.886845 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9svlh\" (UID: \"4f4ac16e-c282-4fa3-bdbe-e5f848782f47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9svlh" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.886887 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9svlh\" (UID: \"4f4ac16e-c282-4fa3-bdbe-e5f848782f47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9svlh" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.886920 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5whjv\" (UniqueName: \"kubernetes.io/projected/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-kube-api-access-5whjv\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9svlh\" (UID: \"4f4ac16e-c282-4fa3-bdbe-e5f848782f47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9svlh" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.886951 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9svlh\" (UID: \"4f4ac16e-c282-4fa3-bdbe-e5f848782f47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9svlh" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.887838 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9svlh\" (UID: \"4f4ac16e-c282-4fa3-bdbe-e5f848782f47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9svlh" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.890062 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9svlh\" (UID: \"4f4ac16e-c282-4fa3-bdbe-e5f848782f47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9svlh" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.890245 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9svlh\" (UID: \"4f4ac16e-c282-4fa3-bdbe-e5f848782f47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9svlh" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.890302 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9svlh\" (UID: \"4f4ac16e-c282-4fa3-bdbe-e5f848782f47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9svlh" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.900354 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5whjv\" (UniqueName: \"kubernetes.io/projected/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-kube-api-access-5whjv\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9svlh\" (UID: \"4f4ac16e-c282-4fa3-bdbe-e5f848782f47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9svlh" Nov 23 07:11:39 crc kubenswrapper[4559]: I1123 07:11:39.926093 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9svlh" Nov 23 07:11:40 crc kubenswrapper[4559]: W1123 07:11:40.346222 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f4ac16e_c282_4fa3_bdbe_e5f848782f47.slice/crio-ff49f4c81a6c56dbdee531199b06064b5733d3361975ea1b2923c3e516202ebb WatchSource:0}: Error finding container ff49f4c81a6c56dbdee531199b06064b5733d3361975ea1b2923c3e516202ebb: Status 404 returned error can't find the container with id ff49f4c81a6c56dbdee531199b06064b5733d3361975ea1b2923c3e516202ebb Nov 23 07:11:40 crc kubenswrapper[4559]: I1123 07:11:40.347107 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-9svlh"] Nov 23 07:11:40 crc kubenswrapper[4559]: I1123 07:11:40.524498 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9svlh" event={"ID":"4f4ac16e-c282-4fa3-bdbe-e5f848782f47","Type":"ContainerStarted","Data":"ff49f4c81a6c56dbdee531199b06064b5733d3361975ea1b2923c3e516202ebb"} Nov 23 07:11:41 crc kubenswrapper[4559]: I1123 07:11:41.533179 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9svlh" event={"ID":"4f4ac16e-c282-4fa3-bdbe-e5f848782f47","Type":"ContainerStarted","Data":"76216dd982130e3fd2346c3649a2b853a8a93fe9d66c962dbe4dc4b30d4ac85c"} Nov 23 07:11:41 crc kubenswrapper[4559]: I1123 07:11:41.550393 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9svlh" podStartSLOduration=1.9769549290000001 podStartE2EDuration="2.550377421s" podCreationTimestamp="2025-11-23 07:11:39 +0000 UTC" firstStartedPulling="2025-11-23 07:11:40.348005996 +0000 UTC m=+1602.369991610" lastFinishedPulling="2025-11-23 07:11:40.921428497 +0000 UTC m=+1602.943414102" observedRunningTime="2025-11-23 07:11:41.544611558 +0000 UTC m=+1603.566597172" watchObservedRunningTime="2025-11-23 07:11:41.550377421 +0000 UTC m=+1603.572363036" Nov 23 07:11:43 crc kubenswrapper[4559]: I1123 07:11:43.690446 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-r7n69" Nov 23 07:11:43 crc kubenswrapper[4559]: I1123 07:11:43.691092 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-r7n69" Nov 23 07:11:43 crc kubenswrapper[4559]: I1123 07:11:43.732752 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-r7n69" Nov 23 07:11:44 crc kubenswrapper[4559]: I1123 07:11:44.583756 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-r7n69" Nov 23 07:11:44 crc kubenswrapper[4559]: I1123 07:11:44.619219 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-r7n69"] Nov 23 07:11:45 crc kubenswrapper[4559]: I1123 07:11:45.274005 4559 scope.go:117] "RemoveContainer" containerID="ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b" Nov 23 07:11:45 crc kubenswrapper[4559]: E1123 07:11:45.274218 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:11:46 crc kubenswrapper[4559]: I1123 07:11:46.565152 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-r7n69" podUID="cc47e6ec-9e05-422c-b0af-08ec3016cd47" containerName="registry-server" containerID="cri-o://104bf0aac053715be0752abf67831d67ceead7eecacfa280fa05220b748c8a47" gracePeriod=2 Nov 23 07:11:46 crc kubenswrapper[4559]: I1123 07:11:46.916011 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r7n69" Nov 23 07:11:47 crc kubenswrapper[4559]: I1123 07:11:47.098018 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc47e6ec-9e05-422c-b0af-08ec3016cd47-utilities\") pod \"cc47e6ec-9e05-422c-b0af-08ec3016cd47\" (UID: \"cc47e6ec-9e05-422c-b0af-08ec3016cd47\") " Nov 23 07:11:47 crc kubenswrapper[4559]: I1123 07:11:47.098230 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc47e6ec-9e05-422c-b0af-08ec3016cd47-catalog-content\") pod \"cc47e6ec-9e05-422c-b0af-08ec3016cd47\" (UID: \"cc47e6ec-9e05-422c-b0af-08ec3016cd47\") " Nov 23 07:11:47 crc kubenswrapper[4559]: I1123 07:11:47.098281 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjf6p\" (UniqueName: \"kubernetes.io/projected/cc47e6ec-9e05-422c-b0af-08ec3016cd47-kube-api-access-kjf6p\") pod \"cc47e6ec-9e05-422c-b0af-08ec3016cd47\" (UID: \"cc47e6ec-9e05-422c-b0af-08ec3016cd47\") " Nov 23 07:11:47 crc kubenswrapper[4559]: I1123 07:11:47.098695 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc47e6ec-9e05-422c-b0af-08ec3016cd47-utilities" (OuterVolumeSpecName: "utilities") pod "cc47e6ec-9e05-422c-b0af-08ec3016cd47" (UID: "cc47e6ec-9e05-422c-b0af-08ec3016cd47"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:11:47 crc kubenswrapper[4559]: I1123 07:11:47.102473 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc47e6ec-9e05-422c-b0af-08ec3016cd47-kube-api-access-kjf6p" (OuterVolumeSpecName: "kube-api-access-kjf6p") pod "cc47e6ec-9e05-422c-b0af-08ec3016cd47" (UID: "cc47e6ec-9e05-422c-b0af-08ec3016cd47"). InnerVolumeSpecName "kube-api-access-kjf6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:11:47 crc kubenswrapper[4559]: I1123 07:11:47.110975 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc47e6ec-9e05-422c-b0af-08ec3016cd47-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cc47e6ec-9e05-422c-b0af-08ec3016cd47" (UID: "cc47e6ec-9e05-422c-b0af-08ec3016cd47"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:11:47 crc kubenswrapper[4559]: I1123 07:11:47.200301 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjf6p\" (UniqueName: \"kubernetes.io/projected/cc47e6ec-9e05-422c-b0af-08ec3016cd47-kube-api-access-kjf6p\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:47 crc kubenswrapper[4559]: I1123 07:11:47.200332 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc47e6ec-9e05-422c-b0af-08ec3016cd47-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:47 crc kubenswrapper[4559]: I1123 07:11:47.200343 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc47e6ec-9e05-422c-b0af-08ec3016cd47-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:11:47 crc kubenswrapper[4559]: I1123 07:11:47.572959 4559 generic.go:334] "Generic (PLEG): container finished" podID="cc47e6ec-9e05-422c-b0af-08ec3016cd47" containerID="104bf0aac053715be0752abf67831d67ceead7eecacfa280fa05220b748c8a47" exitCode=0 Nov 23 07:11:47 crc kubenswrapper[4559]: I1123 07:11:47.572998 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r7n69" Nov 23 07:11:47 crc kubenswrapper[4559]: I1123 07:11:47.572995 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r7n69" event={"ID":"cc47e6ec-9e05-422c-b0af-08ec3016cd47","Type":"ContainerDied","Data":"104bf0aac053715be0752abf67831d67ceead7eecacfa280fa05220b748c8a47"} Nov 23 07:11:47 crc kubenswrapper[4559]: I1123 07:11:47.573141 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r7n69" event={"ID":"cc47e6ec-9e05-422c-b0af-08ec3016cd47","Type":"ContainerDied","Data":"e77df62d71a12b82a7706716d54b917ba4fe672a76b6e06702edae9d0f2fc7ca"} Nov 23 07:11:47 crc kubenswrapper[4559]: I1123 07:11:47.573172 4559 scope.go:117] "RemoveContainer" containerID="104bf0aac053715be0752abf67831d67ceead7eecacfa280fa05220b748c8a47" Nov 23 07:11:47 crc kubenswrapper[4559]: I1123 07:11:47.590598 4559 scope.go:117] "RemoveContainer" containerID="94c4e3a5c2a641bc6fdd936ac6b2309513f9582119d67750cc92c8c9b08064c3" Nov 23 07:11:47 crc kubenswrapper[4559]: I1123 07:11:47.594269 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-r7n69"] Nov 23 07:11:47 crc kubenswrapper[4559]: I1123 07:11:47.600000 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-r7n69"] Nov 23 07:11:47 crc kubenswrapper[4559]: I1123 07:11:47.620841 4559 scope.go:117] "RemoveContainer" containerID="0b8cf628688fe803e8f3376371a3e3f0bfb7b877de4f7ab9aa7c97bfb8ce2a0a" Nov 23 07:11:47 crc kubenswrapper[4559]: I1123 07:11:47.640962 4559 scope.go:117] "RemoveContainer" containerID="104bf0aac053715be0752abf67831d67ceead7eecacfa280fa05220b748c8a47" Nov 23 07:11:47 crc kubenswrapper[4559]: E1123 07:11:47.641357 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"104bf0aac053715be0752abf67831d67ceead7eecacfa280fa05220b748c8a47\": container with ID starting with 104bf0aac053715be0752abf67831d67ceead7eecacfa280fa05220b748c8a47 not found: ID does not exist" containerID="104bf0aac053715be0752abf67831d67ceead7eecacfa280fa05220b748c8a47" Nov 23 07:11:47 crc kubenswrapper[4559]: I1123 07:11:47.641387 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"104bf0aac053715be0752abf67831d67ceead7eecacfa280fa05220b748c8a47"} err="failed to get container status \"104bf0aac053715be0752abf67831d67ceead7eecacfa280fa05220b748c8a47\": rpc error: code = NotFound desc = could not find container \"104bf0aac053715be0752abf67831d67ceead7eecacfa280fa05220b748c8a47\": container with ID starting with 104bf0aac053715be0752abf67831d67ceead7eecacfa280fa05220b748c8a47 not found: ID does not exist" Nov 23 07:11:47 crc kubenswrapper[4559]: I1123 07:11:47.641405 4559 scope.go:117] "RemoveContainer" containerID="94c4e3a5c2a641bc6fdd936ac6b2309513f9582119d67750cc92c8c9b08064c3" Nov 23 07:11:47 crc kubenswrapper[4559]: E1123 07:11:47.641713 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94c4e3a5c2a641bc6fdd936ac6b2309513f9582119d67750cc92c8c9b08064c3\": container with ID starting with 94c4e3a5c2a641bc6fdd936ac6b2309513f9582119d67750cc92c8c9b08064c3 not found: ID does not exist" containerID="94c4e3a5c2a641bc6fdd936ac6b2309513f9582119d67750cc92c8c9b08064c3" Nov 23 07:11:47 crc kubenswrapper[4559]: I1123 07:11:47.641745 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94c4e3a5c2a641bc6fdd936ac6b2309513f9582119d67750cc92c8c9b08064c3"} err="failed to get container status \"94c4e3a5c2a641bc6fdd936ac6b2309513f9582119d67750cc92c8c9b08064c3\": rpc error: code = NotFound desc = could not find container \"94c4e3a5c2a641bc6fdd936ac6b2309513f9582119d67750cc92c8c9b08064c3\": container with ID starting with 94c4e3a5c2a641bc6fdd936ac6b2309513f9582119d67750cc92c8c9b08064c3 not found: ID does not exist" Nov 23 07:11:47 crc kubenswrapper[4559]: I1123 07:11:47.641769 4559 scope.go:117] "RemoveContainer" containerID="0b8cf628688fe803e8f3376371a3e3f0bfb7b877de4f7ab9aa7c97bfb8ce2a0a" Nov 23 07:11:47 crc kubenswrapper[4559]: E1123 07:11:47.642037 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b8cf628688fe803e8f3376371a3e3f0bfb7b877de4f7ab9aa7c97bfb8ce2a0a\": container with ID starting with 0b8cf628688fe803e8f3376371a3e3f0bfb7b877de4f7ab9aa7c97bfb8ce2a0a not found: ID does not exist" containerID="0b8cf628688fe803e8f3376371a3e3f0bfb7b877de4f7ab9aa7c97bfb8ce2a0a" Nov 23 07:11:47 crc kubenswrapper[4559]: I1123 07:11:47.642073 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b8cf628688fe803e8f3376371a3e3f0bfb7b877de4f7ab9aa7c97bfb8ce2a0a"} err="failed to get container status \"0b8cf628688fe803e8f3376371a3e3f0bfb7b877de4f7ab9aa7c97bfb8ce2a0a\": rpc error: code = NotFound desc = could not find container \"0b8cf628688fe803e8f3376371a3e3f0bfb7b877de4f7ab9aa7c97bfb8ce2a0a\": container with ID starting with 0b8cf628688fe803e8f3376371a3e3f0bfb7b877de4f7ab9aa7c97bfb8ce2a0a not found: ID does not exist" Nov 23 07:11:48 crc kubenswrapper[4559]: I1123 07:11:48.282477 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc47e6ec-9e05-422c-b0af-08ec3016cd47" path="/var/lib/kubelet/pods/cc47e6ec-9e05-422c-b0af-08ec3016cd47/volumes" Nov 23 07:11:56 crc kubenswrapper[4559]: I1123 07:11:56.273613 4559 scope.go:117] "RemoveContainer" containerID="ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b" Nov 23 07:11:56 crc kubenswrapper[4559]: E1123 07:11:56.274136 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:12:05 crc kubenswrapper[4559]: I1123 07:12:05.028704 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-msr7f"] Nov 23 07:12:05 crc kubenswrapper[4559]: I1123 07:12:05.035751 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-msr7f"] Nov 23 07:12:06 crc kubenswrapper[4559]: I1123 07:12:06.281533 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c818ed2a-d851-4a13-b67d-967f178654dc" path="/var/lib/kubelet/pods/c818ed2a-d851-4a13-b67d-967f178654dc/volumes" Nov 23 07:12:10 crc kubenswrapper[4559]: I1123 07:12:10.273501 4559 scope.go:117] "RemoveContainer" containerID="ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b" Nov 23 07:12:10 crc kubenswrapper[4559]: E1123 07:12:10.273895 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:12:13 crc kubenswrapper[4559]: I1123 07:12:13.420749 4559 scope.go:117] "RemoveContainer" containerID="56edeb6a2681718e3f9d3ecb52b31a83a292944b5f25164edc1255ba37888f0b" Nov 23 07:12:13 crc kubenswrapper[4559]: I1123 07:12:13.459274 4559 scope.go:117] "RemoveContainer" containerID="ee299c98c771228f28a972b00f9944391ef80d73a6ae4a0822f84f31bcb2b1ed" Nov 23 07:12:13 crc kubenswrapper[4559]: I1123 07:12:13.490829 4559 scope.go:117] "RemoveContainer" containerID="eba4f1b416cd79fd936c7ba6b1670232b97f78c120cfd3d6633857972e6fe662" Nov 23 07:12:22 crc kubenswrapper[4559]: I1123 07:12:22.801611 4559 generic.go:334] "Generic (PLEG): container finished" podID="4f4ac16e-c282-4fa3-bdbe-e5f848782f47" containerID="76216dd982130e3fd2346c3649a2b853a8a93fe9d66c962dbe4dc4b30d4ac85c" exitCode=0 Nov 23 07:12:22 crc kubenswrapper[4559]: I1123 07:12:22.801708 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9svlh" event={"ID":"4f4ac16e-c282-4fa3-bdbe-e5f848782f47","Type":"ContainerDied","Data":"76216dd982130e3fd2346c3649a2b853a8a93fe9d66c962dbe4dc4b30d4ac85c"} Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.091305 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9svlh" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.260517 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-ssh-key\") pod \"4f4ac16e-c282-4fa3-bdbe-e5f848782f47\" (UID: \"4f4ac16e-c282-4fa3-bdbe-e5f848782f47\") " Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.260562 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-ovncontroller-config-0\") pod \"4f4ac16e-c282-4fa3-bdbe-e5f848782f47\" (UID: \"4f4ac16e-c282-4fa3-bdbe-e5f848782f47\") " Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.260670 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-inventory\") pod \"4f4ac16e-c282-4fa3-bdbe-e5f848782f47\" (UID: \"4f4ac16e-c282-4fa3-bdbe-e5f848782f47\") " Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.260692 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5whjv\" (UniqueName: \"kubernetes.io/projected/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-kube-api-access-5whjv\") pod \"4f4ac16e-c282-4fa3-bdbe-e5f848782f47\" (UID: \"4f4ac16e-c282-4fa3-bdbe-e5f848782f47\") " Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.260775 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-ovn-combined-ca-bundle\") pod \"4f4ac16e-c282-4fa3-bdbe-e5f848782f47\" (UID: \"4f4ac16e-c282-4fa3-bdbe-e5f848782f47\") " Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.265771 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-kube-api-access-5whjv" (OuterVolumeSpecName: "kube-api-access-5whjv") pod "4f4ac16e-c282-4fa3-bdbe-e5f848782f47" (UID: "4f4ac16e-c282-4fa3-bdbe-e5f848782f47"). InnerVolumeSpecName "kube-api-access-5whjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.265779 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "4f4ac16e-c282-4fa3-bdbe-e5f848782f47" (UID: "4f4ac16e-c282-4fa3-bdbe-e5f848782f47"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.280323 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "4f4ac16e-c282-4fa3-bdbe-e5f848782f47" (UID: "4f4ac16e-c282-4fa3-bdbe-e5f848782f47"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.280957 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4f4ac16e-c282-4fa3-bdbe-e5f848782f47" (UID: "4f4ac16e-c282-4fa3-bdbe-e5f848782f47"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.281732 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-inventory" (OuterVolumeSpecName: "inventory") pod "4f4ac16e-c282-4fa3-bdbe-e5f848782f47" (UID: "4f4ac16e-c282-4fa3-bdbe-e5f848782f47"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.362481 4559 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.362508 4559 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.362518 4559 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.362526 4559 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.362534 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5whjv\" (UniqueName: \"kubernetes.io/projected/4f4ac16e-c282-4fa3-bdbe-e5f848782f47-kube-api-access-5whjv\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.814662 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9svlh" event={"ID":"4f4ac16e-c282-4fa3-bdbe-e5f848782f47","Type":"ContainerDied","Data":"ff49f4c81a6c56dbdee531199b06064b5733d3361975ea1b2923c3e516202ebb"} Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.814697 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff49f4c81a6c56dbdee531199b06064b5733d3361975ea1b2923c3e516202ebb" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.814703 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9svlh" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.875686 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n"] Nov 23 07:12:24 crc kubenswrapper[4559]: E1123 07:12:24.875975 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f4ac16e-c282-4fa3-bdbe-e5f848782f47" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.875991 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f4ac16e-c282-4fa3-bdbe-e5f848782f47" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 23 07:12:24 crc kubenswrapper[4559]: E1123 07:12:24.876014 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc47e6ec-9e05-422c-b0af-08ec3016cd47" containerName="extract-content" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.876020 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc47e6ec-9e05-422c-b0af-08ec3016cd47" containerName="extract-content" Nov 23 07:12:24 crc kubenswrapper[4559]: E1123 07:12:24.876040 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc47e6ec-9e05-422c-b0af-08ec3016cd47" containerName="registry-server" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.876046 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc47e6ec-9e05-422c-b0af-08ec3016cd47" containerName="registry-server" Nov 23 07:12:24 crc kubenswrapper[4559]: E1123 07:12:24.876059 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc47e6ec-9e05-422c-b0af-08ec3016cd47" containerName="extract-utilities" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.876065 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc47e6ec-9e05-422c-b0af-08ec3016cd47" containerName="extract-utilities" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.876202 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc47e6ec-9e05-422c-b0af-08ec3016cd47" containerName="registry-server" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.876224 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f4ac16e-c282-4fa3-bdbe-e5f848782f47" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.876724 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.879214 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.879499 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mj6c" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.879634 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.879748 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.879778 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.880953 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.893652 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n"] Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.971776 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n\" (UID: \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.971827 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n\" (UID: \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.971937 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n\" (UID: \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.971961 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n\" (UID: \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.971983 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thjz4\" (UniqueName: \"kubernetes.io/projected/4f9b53ba-da26-40bb-9819-cdeb54deaef2-kube-api-access-thjz4\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n\" (UID: \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n" Nov 23 07:12:24 crc kubenswrapper[4559]: I1123 07:12:24.972021 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n\" (UID: \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n" Nov 23 07:12:25 crc kubenswrapper[4559]: I1123 07:12:25.073002 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n\" (UID: \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n" Nov 23 07:12:25 crc kubenswrapper[4559]: I1123 07:12:25.073251 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n\" (UID: \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n" Nov 23 07:12:25 crc kubenswrapper[4559]: I1123 07:12:25.073271 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n\" (UID: \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n" Nov 23 07:12:25 crc kubenswrapper[4559]: I1123 07:12:25.073294 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thjz4\" (UniqueName: \"kubernetes.io/projected/4f9b53ba-da26-40bb-9819-cdeb54deaef2-kube-api-access-thjz4\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n\" (UID: \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n" Nov 23 07:12:25 crc kubenswrapper[4559]: I1123 07:12:25.073328 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n\" (UID: \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n" Nov 23 07:12:25 crc kubenswrapper[4559]: I1123 07:12:25.073499 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n\" (UID: \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n" Nov 23 07:12:25 crc kubenswrapper[4559]: I1123 07:12:25.076055 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n\" (UID: \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n" Nov 23 07:12:25 crc kubenswrapper[4559]: I1123 07:12:25.076138 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n\" (UID: \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n" Nov 23 07:12:25 crc kubenswrapper[4559]: I1123 07:12:25.077027 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n\" (UID: \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n" Nov 23 07:12:25 crc kubenswrapper[4559]: I1123 07:12:25.077365 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n\" (UID: \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n" Nov 23 07:12:25 crc kubenswrapper[4559]: I1123 07:12:25.077535 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n\" (UID: \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n" Nov 23 07:12:25 crc kubenswrapper[4559]: I1123 07:12:25.086529 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thjz4\" (UniqueName: \"kubernetes.io/projected/4f9b53ba-da26-40bb-9819-cdeb54deaef2-kube-api-access-thjz4\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n\" (UID: \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n" Nov 23 07:12:25 crc kubenswrapper[4559]: I1123 07:12:25.190588 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n" Nov 23 07:12:25 crc kubenswrapper[4559]: I1123 07:12:25.273938 4559 scope.go:117] "RemoveContainer" containerID="ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b" Nov 23 07:12:25 crc kubenswrapper[4559]: E1123 07:12:25.274199 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:12:25 crc kubenswrapper[4559]: I1123 07:12:25.598186 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n"] Nov 23 07:12:25 crc kubenswrapper[4559]: I1123 07:12:25.821439 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n" event={"ID":"4f9b53ba-da26-40bb-9819-cdeb54deaef2","Type":"ContainerStarted","Data":"b536c7b62cd2f880c3545c3b6bd751841786d370f39a011fd9a9088978952d77"} Nov 23 07:12:26 crc kubenswrapper[4559]: I1123 07:12:26.829142 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n" event={"ID":"4f9b53ba-da26-40bb-9819-cdeb54deaef2","Type":"ContainerStarted","Data":"7db6ecaf8816063d648c81e59add0b173057ce4d93e8a80f3069bcd69766014e"} Nov 23 07:12:26 crc kubenswrapper[4559]: I1123 07:12:26.841389 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n" podStartSLOduration=2.283361044 podStartE2EDuration="2.841374624s" podCreationTimestamp="2025-11-23 07:12:24 +0000 UTC" firstStartedPulling="2025-11-23 07:12:25.602004117 +0000 UTC m=+1647.623989732" lastFinishedPulling="2025-11-23 07:12:26.160017698 +0000 UTC m=+1648.182003312" observedRunningTime="2025-11-23 07:12:26.840983618 +0000 UTC m=+1648.862969233" watchObservedRunningTime="2025-11-23 07:12:26.841374624 +0000 UTC m=+1648.863360238" Nov 23 07:12:40 crc kubenswrapper[4559]: I1123 07:12:40.274213 4559 scope.go:117] "RemoveContainer" containerID="ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b" Nov 23 07:12:40 crc kubenswrapper[4559]: E1123 07:12:40.274746 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:12:53 crc kubenswrapper[4559]: I1123 07:12:53.274200 4559 scope.go:117] "RemoveContainer" containerID="ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b" Nov 23 07:12:53 crc kubenswrapper[4559]: E1123 07:12:53.274738 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:12:58 crc kubenswrapper[4559]: I1123 07:12:58.034743 4559 generic.go:334] "Generic (PLEG): container finished" podID="4f9b53ba-da26-40bb-9819-cdeb54deaef2" containerID="7db6ecaf8816063d648c81e59add0b173057ce4d93e8a80f3069bcd69766014e" exitCode=0 Nov 23 07:12:58 crc kubenswrapper[4559]: I1123 07:12:58.034816 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n" event={"ID":"4f9b53ba-da26-40bb-9819-cdeb54deaef2","Type":"ContainerDied","Data":"7db6ecaf8816063d648c81e59add0b173057ce4d93e8a80f3069bcd69766014e"} Nov 23 07:12:59 crc kubenswrapper[4559]: I1123 07:12:59.326884 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n" Nov 23 07:12:59 crc kubenswrapper[4559]: I1123 07:12:59.509133 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-nova-metadata-neutron-config-0\") pod \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\" (UID: \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\") " Nov 23 07:12:59 crc kubenswrapper[4559]: I1123 07:12:59.509211 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thjz4\" (UniqueName: \"kubernetes.io/projected/4f9b53ba-da26-40bb-9819-cdeb54deaef2-kube-api-access-thjz4\") pod \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\" (UID: \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\") " Nov 23 07:12:59 crc kubenswrapper[4559]: I1123 07:12:59.509234 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-ssh-key\") pod \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\" (UID: \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\") " Nov 23 07:12:59 crc kubenswrapper[4559]: I1123 07:12:59.509342 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-inventory\") pod \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\" (UID: \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\") " Nov 23 07:12:59 crc kubenswrapper[4559]: I1123 07:12:59.509370 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-neutron-ovn-metadata-agent-neutron-config-0\") pod \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\" (UID: \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\") " Nov 23 07:12:59 crc kubenswrapper[4559]: I1123 07:12:59.509445 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-neutron-metadata-combined-ca-bundle\") pod \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\" (UID: \"4f9b53ba-da26-40bb-9819-cdeb54deaef2\") " Nov 23 07:12:59 crc kubenswrapper[4559]: I1123 07:12:59.513713 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "4f9b53ba-da26-40bb-9819-cdeb54deaef2" (UID: "4f9b53ba-da26-40bb-9819-cdeb54deaef2"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:59 crc kubenswrapper[4559]: I1123 07:12:59.513985 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f9b53ba-da26-40bb-9819-cdeb54deaef2-kube-api-access-thjz4" (OuterVolumeSpecName: "kube-api-access-thjz4") pod "4f9b53ba-da26-40bb-9819-cdeb54deaef2" (UID: "4f9b53ba-da26-40bb-9819-cdeb54deaef2"). InnerVolumeSpecName "kube-api-access-thjz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:12:59 crc kubenswrapper[4559]: I1123 07:12:59.529406 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4f9b53ba-da26-40bb-9819-cdeb54deaef2" (UID: "4f9b53ba-da26-40bb-9819-cdeb54deaef2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:59 crc kubenswrapper[4559]: I1123 07:12:59.529915 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-inventory" (OuterVolumeSpecName: "inventory") pod "4f9b53ba-da26-40bb-9819-cdeb54deaef2" (UID: "4f9b53ba-da26-40bb-9819-cdeb54deaef2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:59 crc kubenswrapper[4559]: I1123 07:12:59.530117 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "4f9b53ba-da26-40bb-9819-cdeb54deaef2" (UID: "4f9b53ba-da26-40bb-9819-cdeb54deaef2"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:59 crc kubenswrapper[4559]: I1123 07:12:59.530824 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "4f9b53ba-da26-40bb-9819-cdeb54deaef2" (UID: "4f9b53ba-da26-40bb-9819-cdeb54deaef2"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:12:59 crc kubenswrapper[4559]: I1123 07:12:59.611765 4559 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:59 crc kubenswrapper[4559]: I1123 07:12:59.611793 4559 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:59 crc kubenswrapper[4559]: I1123 07:12:59.611806 4559 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:59 crc kubenswrapper[4559]: I1123 07:12:59.611817 4559 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:59 crc kubenswrapper[4559]: I1123 07:12:59.611826 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thjz4\" (UniqueName: \"kubernetes.io/projected/4f9b53ba-da26-40bb-9819-cdeb54deaef2-kube-api-access-thjz4\") on node \"crc\" DevicePath \"\"" Nov 23 07:12:59 crc kubenswrapper[4559]: I1123 07:12:59.611835 4559 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f9b53ba-da26-40bb-9819-cdeb54deaef2-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.046638 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n" event={"ID":"4f9b53ba-da26-40bb-9819-cdeb54deaef2","Type":"ContainerDied","Data":"b536c7b62cd2f880c3545c3b6bd751841786d370f39a011fd9a9088978952d77"} Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.046689 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b536c7b62cd2f880c3545c3b6bd751841786d370f39a011fd9a9088978952d77" Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.046696 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n" Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.111013 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-njqpj"] Nov 23 07:13:00 crc kubenswrapper[4559]: E1123 07:13:00.111334 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f9b53ba-da26-40bb-9819-cdeb54deaef2" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.111351 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f9b53ba-da26-40bb-9819-cdeb54deaef2" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.111495 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f9b53ba-da26-40bb-9819-cdeb54deaef2" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.111984 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-njqpj" Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.113656 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.113660 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.113699 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.113725 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.113918 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mj6c" Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.138703 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-njqpj"] Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.219615 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa33c87f-c563-4944-883e-3ed5649b96a5-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-njqpj\" (UID: \"aa33c87f-c563-4944-883e-3ed5649b96a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-njqpj" Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.219671 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa33c87f-c563-4944-883e-3ed5649b96a5-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-njqpj\" (UID: \"aa33c87f-c563-4944-883e-3ed5649b96a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-njqpj" Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.219695 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa33c87f-c563-4944-883e-3ed5649b96a5-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-njqpj\" (UID: \"aa33c87f-c563-4944-883e-3ed5649b96a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-njqpj" Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.219713 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/aa33c87f-c563-4944-883e-3ed5649b96a5-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-njqpj\" (UID: \"aa33c87f-c563-4944-883e-3ed5649b96a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-njqpj" Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.219798 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfb6m\" (UniqueName: \"kubernetes.io/projected/aa33c87f-c563-4944-883e-3ed5649b96a5-kube-api-access-cfb6m\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-njqpj\" (UID: \"aa33c87f-c563-4944-883e-3ed5649b96a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-njqpj" Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.320872 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa33c87f-c563-4944-883e-3ed5649b96a5-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-njqpj\" (UID: \"aa33c87f-c563-4944-883e-3ed5649b96a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-njqpj" Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.320918 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/aa33c87f-c563-4944-883e-3ed5649b96a5-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-njqpj\" (UID: \"aa33c87f-c563-4944-883e-3ed5649b96a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-njqpj" Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.320980 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfb6m\" (UniqueName: \"kubernetes.io/projected/aa33c87f-c563-4944-883e-3ed5649b96a5-kube-api-access-cfb6m\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-njqpj\" (UID: \"aa33c87f-c563-4944-883e-3ed5649b96a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-njqpj" Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.321105 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa33c87f-c563-4944-883e-3ed5649b96a5-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-njqpj\" (UID: \"aa33c87f-c563-4944-883e-3ed5649b96a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-njqpj" Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.321144 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa33c87f-c563-4944-883e-3ed5649b96a5-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-njqpj\" (UID: \"aa33c87f-c563-4944-883e-3ed5649b96a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-njqpj" Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.325207 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa33c87f-c563-4944-883e-3ed5649b96a5-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-njqpj\" (UID: \"aa33c87f-c563-4944-883e-3ed5649b96a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-njqpj" Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.325335 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa33c87f-c563-4944-883e-3ed5649b96a5-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-njqpj\" (UID: \"aa33c87f-c563-4944-883e-3ed5649b96a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-njqpj" Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.326053 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa33c87f-c563-4944-883e-3ed5649b96a5-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-njqpj\" (UID: \"aa33c87f-c563-4944-883e-3ed5649b96a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-njqpj" Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.330539 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/aa33c87f-c563-4944-883e-3ed5649b96a5-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-njqpj\" (UID: \"aa33c87f-c563-4944-883e-3ed5649b96a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-njqpj" Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.334532 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfb6m\" (UniqueName: \"kubernetes.io/projected/aa33c87f-c563-4944-883e-3ed5649b96a5-kube-api-access-cfb6m\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-njqpj\" (UID: \"aa33c87f-c563-4944-883e-3ed5649b96a5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-njqpj" Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.438888 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-njqpj" Nov 23 07:13:00 crc kubenswrapper[4559]: I1123 07:13:00.853689 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-njqpj"] Nov 23 07:13:01 crc kubenswrapper[4559]: I1123 07:13:01.053004 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-njqpj" event={"ID":"aa33c87f-c563-4944-883e-3ed5649b96a5","Type":"ContainerStarted","Data":"c1f5a86a0657b16d5047f2e638dc57ae8671d8ff6e440e41dfd1a9beef76fb05"} Nov 23 07:13:02 crc kubenswrapper[4559]: I1123 07:13:02.061300 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-njqpj" event={"ID":"aa33c87f-c563-4944-883e-3ed5649b96a5","Type":"ContainerStarted","Data":"f25eb7dc696ff72afd7ccc06a690d70a7c331dec9faa06101f2c45a52f0ea5f0"} Nov 23 07:13:02 crc kubenswrapper[4559]: I1123 07:13:02.076192 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-njqpj" podStartSLOduration=1.598551274 podStartE2EDuration="2.076178813s" podCreationTimestamp="2025-11-23 07:13:00 +0000 UTC" firstStartedPulling="2025-11-23 07:13:00.857097432 +0000 UTC m=+1682.879083046" lastFinishedPulling="2025-11-23 07:13:01.334724972 +0000 UTC m=+1683.356710585" observedRunningTime="2025-11-23 07:13:02.071319775 +0000 UTC m=+1684.093305389" watchObservedRunningTime="2025-11-23 07:13:02.076178813 +0000 UTC m=+1684.098164427" Nov 23 07:13:05 crc kubenswrapper[4559]: I1123 07:13:05.274606 4559 scope.go:117] "RemoveContainer" containerID="ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b" Nov 23 07:13:05 crc kubenswrapper[4559]: E1123 07:13:05.275777 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:13:16 crc kubenswrapper[4559]: I1123 07:13:16.275374 4559 scope.go:117] "RemoveContainer" containerID="ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b" Nov 23 07:13:16 crc kubenswrapper[4559]: E1123 07:13:16.275887 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:13:30 crc kubenswrapper[4559]: I1123 07:13:30.274092 4559 scope.go:117] "RemoveContainer" containerID="ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b" Nov 23 07:13:30 crc kubenswrapper[4559]: E1123 07:13:30.274562 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:13:43 crc kubenswrapper[4559]: I1123 07:13:43.273601 4559 scope.go:117] "RemoveContainer" containerID="ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b" Nov 23 07:13:43 crc kubenswrapper[4559]: E1123 07:13:43.286268 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:13:58 crc kubenswrapper[4559]: I1123 07:13:58.278153 4559 scope.go:117] "RemoveContainer" containerID="ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b" Nov 23 07:13:58 crc kubenswrapper[4559]: E1123 07:13:58.278614 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:14:10 crc kubenswrapper[4559]: I1123 07:14:10.275688 4559 scope.go:117] "RemoveContainer" containerID="ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b" Nov 23 07:14:10 crc kubenswrapper[4559]: E1123 07:14:10.278021 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:14:25 crc kubenswrapper[4559]: I1123 07:14:25.274209 4559 scope.go:117] "RemoveContainer" containerID="ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b" Nov 23 07:14:25 crc kubenswrapper[4559]: E1123 07:14:25.275521 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:14:36 crc kubenswrapper[4559]: I1123 07:14:36.273132 4559 scope.go:117] "RemoveContainer" containerID="ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b" Nov 23 07:14:36 crc kubenswrapper[4559]: I1123 07:14:36.685082 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerStarted","Data":"eb8bcd23b2c4b87ac90b1547ee493c8aff582ddd59e852cc40a1f99ebbec9b19"} Nov 23 07:15:00 crc kubenswrapper[4559]: I1123 07:15:00.137906 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398035-vrnbg"] Nov 23 07:15:00 crc kubenswrapper[4559]: I1123 07:15:00.139658 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-vrnbg" Nov 23 07:15:00 crc kubenswrapper[4559]: I1123 07:15:00.141857 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 23 07:15:00 crc kubenswrapper[4559]: I1123 07:15:00.141916 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 23 07:15:00 crc kubenswrapper[4559]: I1123 07:15:00.145085 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398035-vrnbg"] Nov 23 07:15:00 crc kubenswrapper[4559]: I1123 07:15:00.241151 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/178f5509-9d5e-43fa-967c-5398ad5f3f9e-secret-volume\") pod \"collect-profiles-29398035-vrnbg\" (UID: \"178f5509-9d5e-43fa-967c-5398ad5f3f9e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-vrnbg" Nov 23 07:15:00 crc kubenswrapper[4559]: I1123 07:15:00.241240 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqdlp\" (UniqueName: \"kubernetes.io/projected/178f5509-9d5e-43fa-967c-5398ad5f3f9e-kube-api-access-jqdlp\") pod \"collect-profiles-29398035-vrnbg\" (UID: \"178f5509-9d5e-43fa-967c-5398ad5f3f9e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-vrnbg" Nov 23 07:15:00 crc kubenswrapper[4559]: I1123 07:15:00.241307 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/178f5509-9d5e-43fa-967c-5398ad5f3f9e-config-volume\") pod \"collect-profiles-29398035-vrnbg\" (UID: \"178f5509-9d5e-43fa-967c-5398ad5f3f9e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-vrnbg" Nov 23 07:15:00 crc kubenswrapper[4559]: I1123 07:15:00.344223 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/178f5509-9d5e-43fa-967c-5398ad5f3f9e-secret-volume\") pod \"collect-profiles-29398035-vrnbg\" (UID: \"178f5509-9d5e-43fa-967c-5398ad5f3f9e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-vrnbg" Nov 23 07:15:00 crc kubenswrapper[4559]: I1123 07:15:00.344766 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqdlp\" (UniqueName: \"kubernetes.io/projected/178f5509-9d5e-43fa-967c-5398ad5f3f9e-kube-api-access-jqdlp\") pod \"collect-profiles-29398035-vrnbg\" (UID: \"178f5509-9d5e-43fa-967c-5398ad5f3f9e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-vrnbg" Nov 23 07:15:00 crc kubenswrapper[4559]: I1123 07:15:00.344853 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/178f5509-9d5e-43fa-967c-5398ad5f3f9e-config-volume\") pod \"collect-profiles-29398035-vrnbg\" (UID: \"178f5509-9d5e-43fa-967c-5398ad5f3f9e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-vrnbg" Nov 23 07:15:00 crc kubenswrapper[4559]: I1123 07:15:00.345805 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/178f5509-9d5e-43fa-967c-5398ad5f3f9e-config-volume\") pod \"collect-profiles-29398035-vrnbg\" (UID: \"178f5509-9d5e-43fa-967c-5398ad5f3f9e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-vrnbg" Nov 23 07:15:00 crc kubenswrapper[4559]: I1123 07:15:00.352346 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/178f5509-9d5e-43fa-967c-5398ad5f3f9e-secret-volume\") pod \"collect-profiles-29398035-vrnbg\" (UID: \"178f5509-9d5e-43fa-967c-5398ad5f3f9e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-vrnbg" Nov 23 07:15:00 crc kubenswrapper[4559]: I1123 07:15:00.360402 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqdlp\" (UniqueName: \"kubernetes.io/projected/178f5509-9d5e-43fa-967c-5398ad5f3f9e-kube-api-access-jqdlp\") pod \"collect-profiles-29398035-vrnbg\" (UID: \"178f5509-9d5e-43fa-967c-5398ad5f3f9e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-vrnbg" Nov 23 07:15:00 crc kubenswrapper[4559]: I1123 07:15:00.460268 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-vrnbg" Nov 23 07:15:00 crc kubenswrapper[4559]: I1123 07:15:00.955183 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398035-vrnbg"] Nov 23 07:15:01 crc kubenswrapper[4559]: I1123 07:15:01.864866 4559 generic.go:334] "Generic (PLEG): container finished" podID="178f5509-9d5e-43fa-967c-5398ad5f3f9e" containerID="83548213b99973c495f141388d701d90909f0ac4615079d2edd57df7a16e1a01" exitCode=0 Nov 23 07:15:01 crc kubenswrapper[4559]: I1123 07:15:01.865256 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-vrnbg" event={"ID":"178f5509-9d5e-43fa-967c-5398ad5f3f9e","Type":"ContainerDied","Data":"83548213b99973c495f141388d701d90909f0ac4615079d2edd57df7a16e1a01"} Nov 23 07:15:01 crc kubenswrapper[4559]: I1123 07:15:01.865291 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-vrnbg" event={"ID":"178f5509-9d5e-43fa-967c-5398ad5f3f9e","Type":"ContainerStarted","Data":"36cfa509d30cf45edbe431820419453fba4e2d29761762b212bf604a5cb88eb0"} Nov 23 07:15:03 crc kubenswrapper[4559]: I1123 07:15:03.129292 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-vrnbg" Nov 23 07:15:03 crc kubenswrapper[4559]: I1123 07:15:03.211686 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/178f5509-9d5e-43fa-967c-5398ad5f3f9e-config-volume\") pod \"178f5509-9d5e-43fa-967c-5398ad5f3f9e\" (UID: \"178f5509-9d5e-43fa-967c-5398ad5f3f9e\") " Nov 23 07:15:03 crc kubenswrapper[4559]: I1123 07:15:03.211761 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqdlp\" (UniqueName: \"kubernetes.io/projected/178f5509-9d5e-43fa-967c-5398ad5f3f9e-kube-api-access-jqdlp\") pod \"178f5509-9d5e-43fa-967c-5398ad5f3f9e\" (UID: \"178f5509-9d5e-43fa-967c-5398ad5f3f9e\") " Nov 23 07:15:03 crc kubenswrapper[4559]: I1123 07:15:03.211823 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/178f5509-9d5e-43fa-967c-5398ad5f3f9e-secret-volume\") pod \"178f5509-9d5e-43fa-967c-5398ad5f3f9e\" (UID: \"178f5509-9d5e-43fa-967c-5398ad5f3f9e\") " Nov 23 07:15:03 crc kubenswrapper[4559]: I1123 07:15:03.212249 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/178f5509-9d5e-43fa-967c-5398ad5f3f9e-config-volume" (OuterVolumeSpecName: "config-volume") pod "178f5509-9d5e-43fa-967c-5398ad5f3f9e" (UID: "178f5509-9d5e-43fa-967c-5398ad5f3f9e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:15:03 crc kubenswrapper[4559]: I1123 07:15:03.218179 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/178f5509-9d5e-43fa-967c-5398ad5f3f9e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "178f5509-9d5e-43fa-967c-5398ad5f3f9e" (UID: "178f5509-9d5e-43fa-967c-5398ad5f3f9e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:15:03 crc kubenswrapper[4559]: I1123 07:15:03.218241 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/178f5509-9d5e-43fa-967c-5398ad5f3f9e-kube-api-access-jqdlp" (OuterVolumeSpecName: "kube-api-access-jqdlp") pod "178f5509-9d5e-43fa-967c-5398ad5f3f9e" (UID: "178f5509-9d5e-43fa-967c-5398ad5f3f9e"). InnerVolumeSpecName "kube-api-access-jqdlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:15:03 crc kubenswrapper[4559]: I1123 07:15:03.312832 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqdlp\" (UniqueName: \"kubernetes.io/projected/178f5509-9d5e-43fa-967c-5398ad5f3f9e-kube-api-access-jqdlp\") on node \"crc\" DevicePath \"\"" Nov 23 07:15:03 crc kubenswrapper[4559]: I1123 07:15:03.312852 4559 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/178f5509-9d5e-43fa-967c-5398ad5f3f9e-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 23 07:15:03 crc kubenswrapper[4559]: I1123 07:15:03.312861 4559 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/178f5509-9d5e-43fa-967c-5398ad5f3f9e-config-volume\") on node \"crc\" DevicePath \"\"" Nov 23 07:15:03 crc kubenswrapper[4559]: I1123 07:15:03.888208 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-vrnbg" event={"ID":"178f5509-9d5e-43fa-967c-5398ad5f3f9e","Type":"ContainerDied","Data":"36cfa509d30cf45edbe431820419453fba4e2d29761762b212bf604a5cb88eb0"} Nov 23 07:15:03 crc kubenswrapper[4559]: I1123 07:15:03.888736 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36cfa509d30cf45edbe431820419453fba4e2d29761762b212bf604a5cb88eb0" Nov 23 07:15:03 crc kubenswrapper[4559]: I1123 07:15:03.888359 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398035-vrnbg" Nov 23 07:15:08 crc kubenswrapper[4559]: I1123 07:15:08.040663 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jp7zp"] Nov 23 07:15:08 crc kubenswrapper[4559]: E1123 07:15:08.042685 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="178f5509-9d5e-43fa-967c-5398ad5f3f9e" containerName="collect-profiles" Nov 23 07:15:08 crc kubenswrapper[4559]: I1123 07:15:08.042703 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="178f5509-9d5e-43fa-967c-5398ad5f3f9e" containerName="collect-profiles" Nov 23 07:15:08 crc kubenswrapper[4559]: I1123 07:15:08.042887 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="178f5509-9d5e-43fa-967c-5398ad5f3f9e" containerName="collect-profiles" Nov 23 07:15:08 crc kubenswrapper[4559]: I1123 07:15:08.044395 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jp7zp" Nov 23 07:15:08 crc kubenswrapper[4559]: I1123 07:15:08.050103 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jp7zp"] Nov 23 07:15:08 crc kubenswrapper[4559]: I1123 07:15:08.115144 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6497be8d-93be-4538-b10c-228b014b537c-catalog-content\") pod \"redhat-operators-jp7zp\" (UID: \"6497be8d-93be-4538-b10c-228b014b537c\") " pod="openshift-marketplace/redhat-operators-jp7zp" Nov 23 07:15:08 crc kubenswrapper[4559]: I1123 07:15:08.115342 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6497be8d-93be-4538-b10c-228b014b537c-utilities\") pod \"redhat-operators-jp7zp\" (UID: \"6497be8d-93be-4538-b10c-228b014b537c\") " pod="openshift-marketplace/redhat-operators-jp7zp" Nov 23 07:15:08 crc kubenswrapper[4559]: I1123 07:15:08.115512 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5x4dc\" (UniqueName: \"kubernetes.io/projected/6497be8d-93be-4538-b10c-228b014b537c-kube-api-access-5x4dc\") pod \"redhat-operators-jp7zp\" (UID: \"6497be8d-93be-4538-b10c-228b014b537c\") " pod="openshift-marketplace/redhat-operators-jp7zp" Nov 23 07:15:08 crc kubenswrapper[4559]: I1123 07:15:08.218220 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6497be8d-93be-4538-b10c-228b014b537c-catalog-content\") pod \"redhat-operators-jp7zp\" (UID: \"6497be8d-93be-4538-b10c-228b014b537c\") " pod="openshift-marketplace/redhat-operators-jp7zp" Nov 23 07:15:08 crc kubenswrapper[4559]: I1123 07:15:08.218307 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6497be8d-93be-4538-b10c-228b014b537c-utilities\") pod \"redhat-operators-jp7zp\" (UID: \"6497be8d-93be-4538-b10c-228b014b537c\") " pod="openshift-marketplace/redhat-operators-jp7zp" Nov 23 07:15:08 crc kubenswrapper[4559]: I1123 07:15:08.218367 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5x4dc\" (UniqueName: \"kubernetes.io/projected/6497be8d-93be-4538-b10c-228b014b537c-kube-api-access-5x4dc\") pod \"redhat-operators-jp7zp\" (UID: \"6497be8d-93be-4538-b10c-228b014b537c\") " pod="openshift-marketplace/redhat-operators-jp7zp" Nov 23 07:15:08 crc kubenswrapper[4559]: I1123 07:15:08.219194 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6497be8d-93be-4538-b10c-228b014b537c-catalog-content\") pod \"redhat-operators-jp7zp\" (UID: \"6497be8d-93be-4538-b10c-228b014b537c\") " pod="openshift-marketplace/redhat-operators-jp7zp" Nov 23 07:15:08 crc kubenswrapper[4559]: I1123 07:15:08.219423 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6497be8d-93be-4538-b10c-228b014b537c-utilities\") pod \"redhat-operators-jp7zp\" (UID: \"6497be8d-93be-4538-b10c-228b014b537c\") " pod="openshift-marketplace/redhat-operators-jp7zp" Nov 23 07:15:08 crc kubenswrapper[4559]: I1123 07:15:08.236019 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5x4dc\" (UniqueName: \"kubernetes.io/projected/6497be8d-93be-4538-b10c-228b014b537c-kube-api-access-5x4dc\") pod \"redhat-operators-jp7zp\" (UID: \"6497be8d-93be-4538-b10c-228b014b537c\") " pod="openshift-marketplace/redhat-operators-jp7zp" Nov 23 07:15:08 crc kubenswrapper[4559]: I1123 07:15:08.360037 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jp7zp" Nov 23 07:15:08 crc kubenswrapper[4559]: I1123 07:15:08.808560 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jp7zp"] Nov 23 07:15:08 crc kubenswrapper[4559]: W1123 07:15:08.814359 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6497be8d_93be_4538_b10c_228b014b537c.slice/crio-a1e59ca6c5b122602dcb921111a21154a5cd78e5f184120a7dd6e0e7ba4224e5 WatchSource:0}: Error finding container a1e59ca6c5b122602dcb921111a21154a5cd78e5f184120a7dd6e0e7ba4224e5: Status 404 returned error can't find the container with id a1e59ca6c5b122602dcb921111a21154a5cd78e5f184120a7dd6e0e7ba4224e5 Nov 23 07:15:08 crc kubenswrapper[4559]: I1123 07:15:08.927292 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jp7zp" event={"ID":"6497be8d-93be-4538-b10c-228b014b537c","Type":"ContainerStarted","Data":"a1e59ca6c5b122602dcb921111a21154a5cd78e5f184120a7dd6e0e7ba4224e5"} Nov 23 07:15:09 crc kubenswrapper[4559]: I1123 07:15:09.940265 4559 generic.go:334] "Generic (PLEG): container finished" podID="6497be8d-93be-4538-b10c-228b014b537c" containerID="d54b442b3aaa676b709bbeef1b52ac57d347cd7c5e4cdd09cf40a65860107df9" exitCode=0 Nov 23 07:15:09 crc kubenswrapper[4559]: I1123 07:15:09.940314 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jp7zp" event={"ID":"6497be8d-93be-4538-b10c-228b014b537c","Type":"ContainerDied","Data":"d54b442b3aaa676b709bbeef1b52ac57d347cd7c5e4cdd09cf40a65860107df9"} Nov 23 07:15:09 crc kubenswrapper[4559]: I1123 07:15:09.943690 4559 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 07:15:10 crc kubenswrapper[4559]: I1123 07:15:10.973263 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jp7zp" event={"ID":"6497be8d-93be-4538-b10c-228b014b537c","Type":"ContainerStarted","Data":"4cd0e40d08a97e38ff8263fc16a35a8373f3b70f93a7e955b713955632264d47"} Nov 23 07:15:12 crc kubenswrapper[4559]: I1123 07:15:12.993776 4559 generic.go:334] "Generic (PLEG): container finished" podID="6497be8d-93be-4538-b10c-228b014b537c" containerID="4cd0e40d08a97e38ff8263fc16a35a8373f3b70f93a7e955b713955632264d47" exitCode=0 Nov 23 07:15:12 crc kubenswrapper[4559]: I1123 07:15:12.993826 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jp7zp" event={"ID":"6497be8d-93be-4538-b10c-228b014b537c","Type":"ContainerDied","Data":"4cd0e40d08a97e38ff8263fc16a35a8373f3b70f93a7e955b713955632264d47"} Nov 23 07:15:14 crc kubenswrapper[4559]: I1123 07:15:14.002912 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jp7zp" event={"ID":"6497be8d-93be-4538-b10c-228b014b537c","Type":"ContainerStarted","Data":"3c01508f530cfdc146c59ba192c36fa8e98a04a8bd2130fe7dad90a47e8969af"} Nov 23 07:15:14 crc kubenswrapper[4559]: I1123 07:15:14.021712 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jp7zp" podStartSLOduration=2.489712713 podStartE2EDuration="6.02169839s" podCreationTimestamp="2025-11-23 07:15:08 +0000 UTC" firstStartedPulling="2025-11-23 07:15:09.943431417 +0000 UTC m=+1811.965417030" lastFinishedPulling="2025-11-23 07:15:13.475417093 +0000 UTC m=+1815.497402707" observedRunningTime="2025-11-23 07:15:14.017186445 +0000 UTC m=+1816.039172059" watchObservedRunningTime="2025-11-23 07:15:14.02169839 +0000 UTC m=+1816.043684004" Nov 23 07:15:18 crc kubenswrapper[4559]: I1123 07:15:18.360516 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jp7zp" Nov 23 07:15:18 crc kubenswrapper[4559]: I1123 07:15:18.361497 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jp7zp" Nov 23 07:15:18 crc kubenswrapper[4559]: I1123 07:15:18.399204 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jp7zp" Nov 23 07:15:19 crc kubenswrapper[4559]: I1123 07:15:19.080041 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jp7zp" Nov 23 07:15:19 crc kubenswrapper[4559]: I1123 07:15:19.129114 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jp7zp"] Nov 23 07:15:21 crc kubenswrapper[4559]: I1123 07:15:21.064895 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jp7zp" podUID="6497be8d-93be-4538-b10c-228b014b537c" containerName="registry-server" containerID="cri-o://3c01508f530cfdc146c59ba192c36fa8e98a04a8bd2130fe7dad90a47e8969af" gracePeriod=2 Nov 23 07:15:21 crc kubenswrapper[4559]: I1123 07:15:21.503490 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jp7zp" Nov 23 07:15:21 crc kubenswrapper[4559]: I1123 07:15:21.601041 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6497be8d-93be-4538-b10c-228b014b537c-utilities\") pod \"6497be8d-93be-4538-b10c-228b014b537c\" (UID: \"6497be8d-93be-4538-b10c-228b014b537c\") " Nov 23 07:15:21 crc kubenswrapper[4559]: I1123 07:15:21.601200 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6497be8d-93be-4538-b10c-228b014b537c-catalog-content\") pod \"6497be8d-93be-4538-b10c-228b014b537c\" (UID: \"6497be8d-93be-4538-b10c-228b014b537c\") " Nov 23 07:15:21 crc kubenswrapper[4559]: I1123 07:15:21.601262 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5x4dc\" (UniqueName: \"kubernetes.io/projected/6497be8d-93be-4538-b10c-228b014b537c-kube-api-access-5x4dc\") pod \"6497be8d-93be-4538-b10c-228b014b537c\" (UID: \"6497be8d-93be-4538-b10c-228b014b537c\") " Nov 23 07:15:21 crc kubenswrapper[4559]: I1123 07:15:21.601930 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6497be8d-93be-4538-b10c-228b014b537c-utilities" (OuterVolumeSpecName: "utilities") pod "6497be8d-93be-4538-b10c-228b014b537c" (UID: "6497be8d-93be-4538-b10c-228b014b537c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:15:21 crc kubenswrapper[4559]: I1123 07:15:21.606738 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6497be8d-93be-4538-b10c-228b014b537c-kube-api-access-5x4dc" (OuterVolumeSpecName: "kube-api-access-5x4dc") pod "6497be8d-93be-4538-b10c-228b014b537c" (UID: "6497be8d-93be-4538-b10c-228b014b537c"). InnerVolumeSpecName "kube-api-access-5x4dc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:15:21 crc kubenswrapper[4559]: I1123 07:15:21.665616 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6497be8d-93be-4538-b10c-228b014b537c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6497be8d-93be-4538-b10c-228b014b537c" (UID: "6497be8d-93be-4538-b10c-228b014b537c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:15:21 crc kubenswrapper[4559]: I1123 07:15:21.703256 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6497be8d-93be-4538-b10c-228b014b537c-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:15:21 crc kubenswrapper[4559]: I1123 07:15:21.703277 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6497be8d-93be-4538-b10c-228b014b537c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:15:21 crc kubenswrapper[4559]: I1123 07:15:21.703288 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5x4dc\" (UniqueName: \"kubernetes.io/projected/6497be8d-93be-4538-b10c-228b014b537c-kube-api-access-5x4dc\") on node \"crc\" DevicePath \"\"" Nov 23 07:15:22 crc kubenswrapper[4559]: I1123 07:15:22.074995 4559 generic.go:334] "Generic (PLEG): container finished" podID="6497be8d-93be-4538-b10c-228b014b537c" containerID="3c01508f530cfdc146c59ba192c36fa8e98a04a8bd2130fe7dad90a47e8969af" exitCode=0 Nov 23 07:15:22 crc kubenswrapper[4559]: I1123 07:15:22.075041 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jp7zp" event={"ID":"6497be8d-93be-4538-b10c-228b014b537c","Type":"ContainerDied","Data":"3c01508f530cfdc146c59ba192c36fa8e98a04a8bd2130fe7dad90a47e8969af"} Nov 23 07:15:22 crc kubenswrapper[4559]: I1123 07:15:22.075081 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jp7zp" Nov 23 07:15:22 crc kubenswrapper[4559]: I1123 07:15:22.075095 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jp7zp" event={"ID":"6497be8d-93be-4538-b10c-228b014b537c","Type":"ContainerDied","Data":"a1e59ca6c5b122602dcb921111a21154a5cd78e5f184120a7dd6e0e7ba4224e5"} Nov 23 07:15:22 crc kubenswrapper[4559]: I1123 07:15:22.075116 4559 scope.go:117] "RemoveContainer" containerID="3c01508f530cfdc146c59ba192c36fa8e98a04a8bd2130fe7dad90a47e8969af" Nov 23 07:15:22 crc kubenswrapper[4559]: I1123 07:15:22.094201 4559 scope.go:117] "RemoveContainer" containerID="4cd0e40d08a97e38ff8263fc16a35a8373f3b70f93a7e955b713955632264d47" Nov 23 07:15:22 crc kubenswrapper[4559]: I1123 07:15:22.103553 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jp7zp"] Nov 23 07:15:22 crc kubenswrapper[4559]: I1123 07:15:22.110976 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jp7zp"] Nov 23 07:15:22 crc kubenswrapper[4559]: I1123 07:15:22.132449 4559 scope.go:117] "RemoveContainer" containerID="d54b442b3aaa676b709bbeef1b52ac57d347cd7c5e4cdd09cf40a65860107df9" Nov 23 07:15:22 crc kubenswrapper[4559]: I1123 07:15:22.147616 4559 scope.go:117] "RemoveContainer" containerID="3c01508f530cfdc146c59ba192c36fa8e98a04a8bd2130fe7dad90a47e8969af" Nov 23 07:15:22 crc kubenswrapper[4559]: E1123 07:15:22.147992 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c01508f530cfdc146c59ba192c36fa8e98a04a8bd2130fe7dad90a47e8969af\": container with ID starting with 3c01508f530cfdc146c59ba192c36fa8e98a04a8bd2130fe7dad90a47e8969af not found: ID does not exist" containerID="3c01508f530cfdc146c59ba192c36fa8e98a04a8bd2130fe7dad90a47e8969af" Nov 23 07:15:22 crc kubenswrapper[4559]: I1123 07:15:22.148024 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c01508f530cfdc146c59ba192c36fa8e98a04a8bd2130fe7dad90a47e8969af"} err="failed to get container status \"3c01508f530cfdc146c59ba192c36fa8e98a04a8bd2130fe7dad90a47e8969af\": rpc error: code = NotFound desc = could not find container \"3c01508f530cfdc146c59ba192c36fa8e98a04a8bd2130fe7dad90a47e8969af\": container with ID starting with 3c01508f530cfdc146c59ba192c36fa8e98a04a8bd2130fe7dad90a47e8969af not found: ID does not exist" Nov 23 07:15:22 crc kubenswrapper[4559]: I1123 07:15:22.148044 4559 scope.go:117] "RemoveContainer" containerID="4cd0e40d08a97e38ff8263fc16a35a8373f3b70f93a7e955b713955632264d47" Nov 23 07:15:22 crc kubenswrapper[4559]: E1123 07:15:22.148257 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cd0e40d08a97e38ff8263fc16a35a8373f3b70f93a7e955b713955632264d47\": container with ID starting with 4cd0e40d08a97e38ff8263fc16a35a8373f3b70f93a7e955b713955632264d47 not found: ID does not exist" containerID="4cd0e40d08a97e38ff8263fc16a35a8373f3b70f93a7e955b713955632264d47" Nov 23 07:15:22 crc kubenswrapper[4559]: I1123 07:15:22.148279 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cd0e40d08a97e38ff8263fc16a35a8373f3b70f93a7e955b713955632264d47"} err="failed to get container status \"4cd0e40d08a97e38ff8263fc16a35a8373f3b70f93a7e955b713955632264d47\": rpc error: code = NotFound desc = could not find container \"4cd0e40d08a97e38ff8263fc16a35a8373f3b70f93a7e955b713955632264d47\": container with ID starting with 4cd0e40d08a97e38ff8263fc16a35a8373f3b70f93a7e955b713955632264d47 not found: ID does not exist" Nov 23 07:15:22 crc kubenswrapper[4559]: I1123 07:15:22.148291 4559 scope.go:117] "RemoveContainer" containerID="d54b442b3aaa676b709bbeef1b52ac57d347cd7c5e4cdd09cf40a65860107df9" Nov 23 07:15:22 crc kubenswrapper[4559]: E1123 07:15:22.148561 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d54b442b3aaa676b709bbeef1b52ac57d347cd7c5e4cdd09cf40a65860107df9\": container with ID starting with d54b442b3aaa676b709bbeef1b52ac57d347cd7c5e4cdd09cf40a65860107df9 not found: ID does not exist" containerID="d54b442b3aaa676b709bbeef1b52ac57d347cd7c5e4cdd09cf40a65860107df9" Nov 23 07:15:22 crc kubenswrapper[4559]: I1123 07:15:22.148586 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d54b442b3aaa676b709bbeef1b52ac57d347cd7c5e4cdd09cf40a65860107df9"} err="failed to get container status \"d54b442b3aaa676b709bbeef1b52ac57d347cd7c5e4cdd09cf40a65860107df9\": rpc error: code = NotFound desc = could not find container \"d54b442b3aaa676b709bbeef1b52ac57d347cd7c5e4cdd09cf40a65860107df9\": container with ID starting with d54b442b3aaa676b709bbeef1b52ac57d347cd7c5e4cdd09cf40a65860107df9 not found: ID does not exist" Nov 23 07:15:22 crc kubenswrapper[4559]: I1123 07:15:22.281923 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6497be8d-93be-4538-b10c-228b014b537c" path="/var/lib/kubelet/pods/6497be8d-93be-4538-b10c-228b014b537c/volumes" Nov 23 07:15:50 crc kubenswrapper[4559]: I1123 07:15:50.283695 4559 generic.go:334] "Generic (PLEG): container finished" podID="aa33c87f-c563-4944-883e-3ed5649b96a5" containerID="f25eb7dc696ff72afd7ccc06a690d70a7c331dec9faa06101f2c45a52f0ea5f0" exitCode=0 Nov 23 07:15:50 crc kubenswrapper[4559]: I1123 07:15:50.283778 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-njqpj" event={"ID":"aa33c87f-c563-4944-883e-3ed5649b96a5","Type":"ContainerDied","Data":"f25eb7dc696ff72afd7ccc06a690d70a7c331dec9faa06101f2c45a52f0ea5f0"} Nov 23 07:15:51 crc kubenswrapper[4559]: I1123 07:15:51.666223 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-njqpj" Nov 23 07:15:51 crc kubenswrapper[4559]: I1123 07:15:51.763272 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa33c87f-c563-4944-883e-3ed5649b96a5-inventory\") pod \"aa33c87f-c563-4944-883e-3ed5649b96a5\" (UID: \"aa33c87f-c563-4944-883e-3ed5649b96a5\") " Nov 23 07:15:51 crc kubenswrapper[4559]: I1123 07:15:51.763321 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa33c87f-c563-4944-883e-3ed5649b96a5-libvirt-combined-ca-bundle\") pod \"aa33c87f-c563-4944-883e-3ed5649b96a5\" (UID: \"aa33c87f-c563-4944-883e-3ed5649b96a5\") " Nov 23 07:15:51 crc kubenswrapper[4559]: I1123 07:15:51.763452 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa33c87f-c563-4944-883e-3ed5649b96a5-ssh-key\") pod \"aa33c87f-c563-4944-883e-3ed5649b96a5\" (UID: \"aa33c87f-c563-4944-883e-3ed5649b96a5\") " Nov 23 07:15:51 crc kubenswrapper[4559]: I1123 07:15:51.763472 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/aa33c87f-c563-4944-883e-3ed5649b96a5-libvirt-secret-0\") pod \"aa33c87f-c563-4944-883e-3ed5649b96a5\" (UID: \"aa33c87f-c563-4944-883e-3ed5649b96a5\") " Nov 23 07:15:51 crc kubenswrapper[4559]: I1123 07:15:51.763543 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfb6m\" (UniqueName: \"kubernetes.io/projected/aa33c87f-c563-4944-883e-3ed5649b96a5-kube-api-access-cfb6m\") pod \"aa33c87f-c563-4944-883e-3ed5649b96a5\" (UID: \"aa33c87f-c563-4944-883e-3ed5649b96a5\") " Nov 23 07:15:51 crc kubenswrapper[4559]: I1123 07:15:51.767610 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa33c87f-c563-4944-883e-3ed5649b96a5-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "aa33c87f-c563-4944-883e-3ed5649b96a5" (UID: "aa33c87f-c563-4944-883e-3ed5649b96a5"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:15:51 crc kubenswrapper[4559]: I1123 07:15:51.767605 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa33c87f-c563-4944-883e-3ed5649b96a5-kube-api-access-cfb6m" (OuterVolumeSpecName: "kube-api-access-cfb6m") pod "aa33c87f-c563-4944-883e-3ed5649b96a5" (UID: "aa33c87f-c563-4944-883e-3ed5649b96a5"). InnerVolumeSpecName "kube-api-access-cfb6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:15:51 crc kubenswrapper[4559]: I1123 07:15:51.783159 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa33c87f-c563-4944-883e-3ed5649b96a5-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "aa33c87f-c563-4944-883e-3ed5649b96a5" (UID: "aa33c87f-c563-4944-883e-3ed5649b96a5"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:15:51 crc kubenswrapper[4559]: I1123 07:15:51.783493 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa33c87f-c563-4944-883e-3ed5649b96a5-inventory" (OuterVolumeSpecName: "inventory") pod "aa33c87f-c563-4944-883e-3ed5649b96a5" (UID: "aa33c87f-c563-4944-883e-3ed5649b96a5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:15:51 crc kubenswrapper[4559]: I1123 07:15:51.784707 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa33c87f-c563-4944-883e-3ed5649b96a5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "aa33c87f-c563-4944-883e-3ed5649b96a5" (UID: "aa33c87f-c563-4944-883e-3ed5649b96a5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:15:51 crc kubenswrapper[4559]: I1123 07:15:51.866234 4559 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/aa33c87f-c563-4944-883e-3ed5649b96a5-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 23 07:15:51 crc kubenswrapper[4559]: I1123 07:15:51.866261 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfb6m\" (UniqueName: \"kubernetes.io/projected/aa33c87f-c563-4944-883e-3ed5649b96a5-kube-api-access-cfb6m\") on node \"crc\" DevicePath \"\"" Nov 23 07:15:51 crc kubenswrapper[4559]: I1123 07:15:51.866273 4559 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa33c87f-c563-4944-883e-3ed5649b96a5-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 07:15:51 crc kubenswrapper[4559]: I1123 07:15:51.866282 4559 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa33c87f-c563-4944-883e-3ed5649b96a5-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:15:51 crc kubenswrapper[4559]: I1123 07:15:51.866290 4559 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa33c87f-c563-4944-883e-3ed5649b96a5-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.300006 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-njqpj" event={"ID":"aa33c87f-c563-4944-883e-3ed5649b96a5","Type":"ContainerDied","Data":"c1f5a86a0657b16d5047f2e638dc57ae8671d8ff6e440e41dfd1a9beef76fb05"} Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.300040 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1f5a86a0657b16d5047f2e638dc57ae8671d8ff6e440e41dfd1a9beef76fb05" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.300063 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-njqpj" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.365760 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h"] Nov 23 07:15:52 crc kubenswrapper[4559]: E1123 07:15:52.366215 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6497be8d-93be-4538-b10c-228b014b537c" containerName="registry-server" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.366235 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="6497be8d-93be-4538-b10c-228b014b537c" containerName="registry-server" Nov 23 07:15:52 crc kubenswrapper[4559]: E1123 07:15:52.366248 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa33c87f-c563-4944-883e-3ed5649b96a5" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.366255 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa33c87f-c563-4944-883e-3ed5649b96a5" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 23 07:15:52 crc kubenswrapper[4559]: E1123 07:15:52.366275 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6497be8d-93be-4538-b10c-228b014b537c" containerName="extract-content" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.366281 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="6497be8d-93be-4538-b10c-228b014b537c" containerName="extract-content" Nov 23 07:15:52 crc kubenswrapper[4559]: E1123 07:15:52.366294 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6497be8d-93be-4538-b10c-228b014b537c" containerName="extract-utilities" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.366299 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="6497be8d-93be-4538-b10c-228b014b537c" containerName="extract-utilities" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.366504 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="6497be8d-93be-4538-b10c-228b014b537c" containerName="registry-server" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.366523 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa33c87f-c563-4944-883e-3ed5649b96a5" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.367200 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.369207 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.369413 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.369955 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.370654 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.371834 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mj6c" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.372154 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.372335 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.384005 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h"] Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.474600 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dfw5h\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.474636 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dfw5h\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.474749 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dfw5h\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.474820 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dfw5h\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.474883 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dfw5h\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.474907 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dfw5h\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.474980 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjnmg\" (UniqueName: \"kubernetes.io/projected/2505b564-8639-464b-b223-91af7ab3661d-kube-api-access-jjnmg\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dfw5h\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.475028 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2505b564-8639-464b-b223-91af7ab3661d-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dfw5h\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.475059 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dfw5h\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.575878 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dfw5h\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.576322 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dfw5h\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.576354 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dfw5h\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.576387 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjnmg\" (UniqueName: \"kubernetes.io/projected/2505b564-8639-464b-b223-91af7ab3661d-kube-api-access-jjnmg\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dfw5h\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.576412 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2505b564-8639-464b-b223-91af7ab3661d-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dfw5h\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.576471 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dfw5h\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.576506 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dfw5h\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.576524 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dfw5h\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.576562 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dfw5h\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.577523 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2505b564-8639-464b-b223-91af7ab3661d-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dfw5h\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.579136 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dfw5h\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.579144 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dfw5h\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.579574 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dfw5h\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.579901 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dfw5h\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.580087 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dfw5h\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.580380 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dfw5h\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.580382 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dfw5h\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.590086 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjnmg\" (UniqueName: \"kubernetes.io/projected/2505b564-8639-464b-b223-91af7ab3661d-kube-api-access-jjnmg\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dfw5h\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:52 crc kubenswrapper[4559]: I1123 07:15:52.680247 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:15:53 crc kubenswrapper[4559]: I1123 07:15:53.097465 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h"] Nov 23 07:15:53 crc kubenswrapper[4559]: I1123 07:15:53.309195 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" event={"ID":"2505b564-8639-464b-b223-91af7ab3661d","Type":"ContainerStarted","Data":"ec9ed69f4aef433bbd11150ce03caa14c8961ed505958764a5fedf391e9c6a92"} Nov 23 07:15:54 crc kubenswrapper[4559]: I1123 07:15:54.318520 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" event={"ID":"2505b564-8639-464b-b223-91af7ab3661d","Type":"ContainerStarted","Data":"d48c03c8216ba3a530282ce4abc6350661535db5b4d3131c9e87f7e3ccc5c21d"} Nov 23 07:15:54 crc kubenswrapper[4559]: I1123 07:15:54.335351 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" podStartSLOduration=1.7698411649999999 podStartE2EDuration="2.335334862s" podCreationTimestamp="2025-11-23 07:15:52 +0000 UTC" firstStartedPulling="2025-11-23 07:15:53.10307669 +0000 UTC m=+1855.125062305" lastFinishedPulling="2025-11-23 07:15:53.668570388 +0000 UTC m=+1855.690556002" observedRunningTime="2025-11-23 07:15:54.331019406 +0000 UTC m=+1856.353005020" watchObservedRunningTime="2025-11-23 07:15:54.335334862 +0000 UTC m=+1856.357320476" Nov 23 07:16:56 crc kubenswrapper[4559]: I1123 07:16:56.167123 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:16:56 crc kubenswrapper[4559]: I1123 07:16:56.167499 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:17:26 crc kubenswrapper[4559]: I1123 07:17:26.166476 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:17:26 crc kubenswrapper[4559]: I1123 07:17:26.166889 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:17:44 crc kubenswrapper[4559]: I1123 07:17:44.123000 4559 generic.go:334] "Generic (PLEG): container finished" podID="2505b564-8639-464b-b223-91af7ab3661d" containerID="d48c03c8216ba3a530282ce4abc6350661535db5b4d3131c9e87f7e3ccc5c21d" exitCode=0 Nov 23 07:17:44 crc kubenswrapper[4559]: I1123 07:17:44.123090 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" event={"ID":"2505b564-8639-464b-b223-91af7ab3661d","Type":"ContainerDied","Data":"d48c03c8216ba3a530282ce4abc6350661535db5b4d3131c9e87f7e3ccc5c21d"} Nov 23 07:17:45 crc kubenswrapper[4559]: I1123 07:17:45.442951 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:17:45 crc kubenswrapper[4559]: I1123 07:17:45.494672 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-combined-ca-bundle\") pod \"2505b564-8639-464b-b223-91af7ab3661d\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " Nov 23 07:17:45 crc kubenswrapper[4559]: I1123 07:17:45.494785 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2505b564-8639-464b-b223-91af7ab3661d-nova-extra-config-0\") pod \"2505b564-8639-464b-b223-91af7ab3661d\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " Nov 23 07:17:45 crc kubenswrapper[4559]: I1123 07:17:45.494918 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-cell1-compute-config-1\") pod \"2505b564-8639-464b-b223-91af7ab3661d\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " Nov 23 07:17:45 crc kubenswrapper[4559]: I1123 07:17:45.495483 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-inventory\") pod \"2505b564-8639-464b-b223-91af7ab3661d\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " Nov 23 07:17:45 crc kubenswrapper[4559]: I1123 07:17:45.495524 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjnmg\" (UniqueName: \"kubernetes.io/projected/2505b564-8639-464b-b223-91af7ab3661d-kube-api-access-jjnmg\") pod \"2505b564-8639-464b-b223-91af7ab3661d\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " Nov 23 07:17:45 crc kubenswrapper[4559]: I1123 07:17:45.495599 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-migration-ssh-key-1\") pod \"2505b564-8639-464b-b223-91af7ab3661d\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " Nov 23 07:17:45 crc kubenswrapper[4559]: I1123 07:17:45.495624 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-ssh-key\") pod \"2505b564-8639-464b-b223-91af7ab3661d\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " Nov 23 07:17:45 crc kubenswrapper[4559]: I1123 07:17:45.496077 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-migration-ssh-key-0\") pod \"2505b564-8639-464b-b223-91af7ab3661d\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " Nov 23 07:17:45 crc kubenswrapper[4559]: I1123 07:17:45.496101 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-cell1-compute-config-0\") pod \"2505b564-8639-464b-b223-91af7ab3661d\" (UID: \"2505b564-8639-464b-b223-91af7ab3661d\") " Nov 23 07:17:45 crc kubenswrapper[4559]: I1123 07:17:45.500040 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "2505b564-8639-464b-b223-91af7ab3661d" (UID: "2505b564-8639-464b-b223-91af7ab3661d"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:17:45 crc kubenswrapper[4559]: I1123 07:17:45.500499 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2505b564-8639-464b-b223-91af7ab3661d-kube-api-access-jjnmg" (OuterVolumeSpecName: "kube-api-access-jjnmg") pod "2505b564-8639-464b-b223-91af7ab3661d" (UID: "2505b564-8639-464b-b223-91af7ab3661d"). InnerVolumeSpecName "kube-api-access-jjnmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:17:45 crc kubenswrapper[4559]: I1123 07:17:45.516711 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2505b564-8639-464b-b223-91af7ab3661d-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "2505b564-8639-464b-b223-91af7ab3661d" (UID: "2505b564-8639-464b-b223-91af7ab3661d"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:17:45 crc kubenswrapper[4559]: I1123 07:17:45.518389 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "2505b564-8639-464b-b223-91af7ab3661d" (UID: "2505b564-8639-464b-b223-91af7ab3661d"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:17:45 crc kubenswrapper[4559]: I1123 07:17:45.518801 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "2505b564-8639-464b-b223-91af7ab3661d" (UID: "2505b564-8639-464b-b223-91af7ab3661d"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:17:45 crc kubenswrapper[4559]: I1123 07:17:45.521244 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "2505b564-8639-464b-b223-91af7ab3661d" (UID: "2505b564-8639-464b-b223-91af7ab3661d"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:17:45 crc kubenswrapper[4559]: I1123 07:17:45.521784 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2505b564-8639-464b-b223-91af7ab3661d" (UID: "2505b564-8639-464b-b223-91af7ab3661d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:17:45 crc kubenswrapper[4559]: I1123 07:17:45.521913 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-inventory" (OuterVolumeSpecName: "inventory") pod "2505b564-8639-464b-b223-91af7ab3661d" (UID: "2505b564-8639-464b-b223-91af7ab3661d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:17:45 crc kubenswrapper[4559]: I1123 07:17:45.522197 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "2505b564-8639-464b-b223-91af7ab3661d" (UID: "2505b564-8639-464b-b223-91af7ab3661d"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:17:45 crc kubenswrapper[4559]: I1123 07:17:45.598712 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjnmg\" (UniqueName: \"kubernetes.io/projected/2505b564-8639-464b-b223-91af7ab3661d-kube-api-access-jjnmg\") on node \"crc\" DevicePath \"\"" Nov 23 07:17:45 crc kubenswrapper[4559]: I1123 07:17:45.598736 4559 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 23 07:17:45 crc kubenswrapper[4559]: I1123 07:17:45.598746 4559 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 07:17:45 crc kubenswrapper[4559]: I1123 07:17:45.598755 4559 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 23 07:17:45 crc kubenswrapper[4559]: I1123 07:17:45.598763 4559 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 23 07:17:45 crc kubenswrapper[4559]: I1123 07:17:45.598771 4559 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:17:45 crc kubenswrapper[4559]: I1123 07:17:45.598782 4559 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2505b564-8639-464b-b223-91af7ab3661d-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Nov 23 07:17:45 crc kubenswrapper[4559]: I1123 07:17:45.598790 4559 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 23 07:17:45 crc kubenswrapper[4559]: I1123 07:17:45.598797 4559 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2505b564-8639-464b-b223-91af7ab3661d-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.138674 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" event={"ID":"2505b564-8639-464b-b223-91af7ab3661d","Type":"ContainerDied","Data":"ec9ed69f4aef433bbd11150ce03caa14c8961ed505958764a5fedf391e9c6a92"} Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.138906 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec9ed69f4aef433bbd11150ce03caa14c8961ed505958764a5fedf391e9c6a92" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.138719 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dfw5h" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.205559 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8"] Nov 23 07:17:46 crc kubenswrapper[4559]: E1123 07:17:46.205910 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2505b564-8639-464b-b223-91af7ab3661d" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.205928 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="2505b564-8639-464b-b223-91af7ab3661d" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.206096 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="2505b564-8639-464b-b223-91af7ab3661d" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.206689 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.208933 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.209064 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.209194 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mj6c" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.209571 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.209871 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.214166 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8"] Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.315129 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8\" (UID: \"14f31bb2-9f00-4931-a84a-900401189cc0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.315437 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8\" (UID: \"14f31bb2-9f00-4931-a84a-900401189cc0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.315587 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8\" (UID: \"14f31bb2-9f00-4931-a84a-900401189cc0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.315737 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8\" (UID: \"14f31bb2-9f00-4931-a84a-900401189cc0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.315939 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xxcz\" (UniqueName: \"kubernetes.io/projected/14f31bb2-9f00-4931-a84a-900401189cc0-kube-api-access-2xxcz\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8\" (UID: \"14f31bb2-9f00-4931-a84a-900401189cc0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.315989 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8\" (UID: \"14f31bb2-9f00-4931-a84a-900401189cc0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.316094 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8\" (UID: \"14f31bb2-9f00-4931-a84a-900401189cc0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.417027 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xxcz\" (UniqueName: \"kubernetes.io/projected/14f31bb2-9f00-4931-a84a-900401189cc0-kube-api-access-2xxcz\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8\" (UID: \"14f31bb2-9f00-4931-a84a-900401189cc0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.417069 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8\" (UID: \"14f31bb2-9f00-4931-a84a-900401189cc0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.417096 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8\" (UID: \"14f31bb2-9f00-4931-a84a-900401189cc0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.417124 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8\" (UID: \"14f31bb2-9f00-4931-a84a-900401189cc0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.417259 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8\" (UID: \"14f31bb2-9f00-4931-a84a-900401189cc0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.417302 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8\" (UID: \"14f31bb2-9f00-4931-a84a-900401189cc0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.417338 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8\" (UID: \"14f31bb2-9f00-4931-a84a-900401189cc0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.420440 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8\" (UID: \"14f31bb2-9f00-4931-a84a-900401189cc0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.420467 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8\" (UID: \"14f31bb2-9f00-4931-a84a-900401189cc0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.420688 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8\" (UID: \"14f31bb2-9f00-4931-a84a-900401189cc0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.421541 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8\" (UID: \"14f31bb2-9f00-4931-a84a-900401189cc0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.428673 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8\" (UID: \"14f31bb2-9f00-4931-a84a-900401189cc0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.429285 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xxcz\" (UniqueName: \"kubernetes.io/projected/14f31bb2-9f00-4931-a84a-900401189cc0-kube-api-access-2xxcz\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8\" (UID: \"14f31bb2-9f00-4931-a84a-900401189cc0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.429286 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8\" (UID: \"14f31bb2-9f00-4931-a84a-900401189cc0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.520491 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" Nov 23 07:17:46 crc kubenswrapper[4559]: I1123 07:17:46.941195 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8"] Nov 23 07:17:47 crc kubenswrapper[4559]: I1123 07:17:47.145960 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" event={"ID":"14f31bb2-9f00-4931-a84a-900401189cc0","Type":"ContainerStarted","Data":"aa7dcc04a8a566fd9fe384e790072cadb04fe9cda9f544ab2b133a36826fb33e"} Nov 23 07:17:48 crc kubenswrapper[4559]: I1123 07:17:48.154243 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" event={"ID":"14f31bb2-9f00-4931-a84a-900401189cc0","Type":"ContainerStarted","Data":"271234e32a4d3ae4d5cb127190f0ba0b823982ecfc2dddadd31f27d098dea80e"} Nov 23 07:17:48 crc kubenswrapper[4559]: I1123 07:17:48.167797 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" podStartSLOduration=1.6618669659999998 podStartE2EDuration="2.167784107s" podCreationTimestamp="2025-11-23 07:17:46 +0000 UTC" firstStartedPulling="2025-11-23 07:17:46.946209042 +0000 UTC m=+1968.968194656" lastFinishedPulling="2025-11-23 07:17:47.452126183 +0000 UTC m=+1969.474111797" observedRunningTime="2025-11-23 07:17:48.166354057 +0000 UTC m=+1970.188339672" watchObservedRunningTime="2025-11-23 07:17:48.167784107 +0000 UTC m=+1970.189769721" Nov 23 07:17:56 crc kubenswrapper[4559]: I1123 07:17:56.166586 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:17:56 crc kubenswrapper[4559]: I1123 07:17:56.167551 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:17:56 crc kubenswrapper[4559]: I1123 07:17:56.167623 4559 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" Nov 23 07:17:56 crc kubenswrapper[4559]: I1123 07:17:56.168123 4559 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"eb8bcd23b2c4b87ac90b1547ee493c8aff582ddd59e852cc40a1f99ebbec9b19"} pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 07:17:56 crc kubenswrapper[4559]: I1123 07:17:56.168182 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" containerID="cri-o://eb8bcd23b2c4b87ac90b1547ee493c8aff582ddd59e852cc40a1f99ebbec9b19" gracePeriod=600 Nov 23 07:17:57 crc kubenswrapper[4559]: I1123 07:17:57.216705 4559 generic.go:334] "Generic (PLEG): container finished" podID="4731beee-0cac-4189-8a70-743b0b709095" containerID="eb8bcd23b2c4b87ac90b1547ee493c8aff582ddd59e852cc40a1f99ebbec9b19" exitCode=0 Nov 23 07:17:57 crc kubenswrapper[4559]: I1123 07:17:57.216908 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerDied","Data":"eb8bcd23b2c4b87ac90b1547ee493c8aff582ddd59e852cc40a1f99ebbec9b19"} Nov 23 07:17:57 crc kubenswrapper[4559]: I1123 07:17:57.217052 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerStarted","Data":"e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f"} Nov 23 07:17:57 crc kubenswrapper[4559]: I1123 07:17:57.217077 4559 scope.go:117] "RemoveContainer" containerID="ae05a823a4b47648498910a3547202153e0673a1e8dcc1300c87482b561ab18b" Nov 23 07:19:15 crc kubenswrapper[4559]: I1123 07:19:15.352334 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tpxl4"] Nov 23 07:19:15 crc kubenswrapper[4559]: I1123 07:19:15.355116 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tpxl4" Nov 23 07:19:15 crc kubenswrapper[4559]: I1123 07:19:15.356557 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b08ab74f-9543-4016-8d4d-00134d2bfe5a-catalog-content\") pod \"certified-operators-tpxl4\" (UID: \"b08ab74f-9543-4016-8d4d-00134d2bfe5a\") " pod="openshift-marketplace/certified-operators-tpxl4" Nov 23 07:19:15 crc kubenswrapper[4559]: I1123 07:19:15.357352 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpjkj\" (UniqueName: \"kubernetes.io/projected/b08ab74f-9543-4016-8d4d-00134d2bfe5a-kube-api-access-wpjkj\") pod \"certified-operators-tpxl4\" (UID: \"b08ab74f-9543-4016-8d4d-00134d2bfe5a\") " pod="openshift-marketplace/certified-operators-tpxl4" Nov 23 07:19:15 crc kubenswrapper[4559]: I1123 07:19:15.357530 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b08ab74f-9543-4016-8d4d-00134d2bfe5a-utilities\") pod \"certified-operators-tpxl4\" (UID: \"b08ab74f-9543-4016-8d4d-00134d2bfe5a\") " pod="openshift-marketplace/certified-operators-tpxl4" Nov 23 07:19:15 crc kubenswrapper[4559]: I1123 07:19:15.361533 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tpxl4"] Nov 23 07:19:15 crc kubenswrapper[4559]: I1123 07:19:15.459246 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b08ab74f-9543-4016-8d4d-00134d2bfe5a-catalog-content\") pod \"certified-operators-tpxl4\" (UID: \"b08ab74f-9543-4016-8d4d-00134d2bfe5a\") " pod="openshift-marketplace/certified-operators-tpxl4" Nov 23 07:19:15 crc kubenswrapper[4559]: I1123 07:19:15.459455 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpjkj\" (UniqueName: \"kubernetes.io/projected/b08ab74f-9543-4016-8d4d-00134d2bfe5a-kube-api-access-wpjkj\") pod \"certified-operators-tpxl4\" (UID: \"b08ab74f-9543-4016-8d4d-00134d2bfe5a\") " pod="openshift-marketplace/certified-operators-tpxl4" Nov 23 07:19:15 crc kubenswrapper[4559]: I1123 07:19:15.459556 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b08ab74f-9543-4016-8d4d-00134d2bfe5a-utilities\") pod \"certified-operators-tpxl4\" (UID: \"b08ab74f-9543-4016-8d4d-00134d2bfe5a\") " pod="openshift-marketplace/certified-operators-tpxl4" Nov 23 07:19:15 crc kubenswrapper[4559]: I1123 07:19:15.459738 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b08ab74f-9543-4016-8d4d-00134d2bfe5a-catalog-content\") pod \"certified-operators-tpxl4\" (UID: \"b08ab74f-9543-4016-8d4d-00134d2bfe5a\") " pod="openshift-marketplace/certified-operators-tpxl4" Nov 23 07:19:15 crc kubenswrapper[4559]: I1123 07:19:15.459905 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b08ab74f-9543-4016-8d4d-00134d2bfe5a-utilities\") pod \"certified-operators-tpxl4\" (UID: \"b08ab74f-9543-4016-8d4d-00134d2bfe5a\") " pod="openshift-marketplace/certified-operators-tpxl4" Nov 23 07:19:15 crc kubenswrapper[4559]: I1123 07:19:15.476222 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpjkj\" (UniqueName: \"kubernetes.io/projected/b08ab74f-9543-4016-8d4d-00134d2bfe5a-kube-api-access-wpjkj\") pod \"certified-operators-tpxl4\" (UID: \"b08ab74f-9543-4016-8d4d-00134d2bfe5a\") " pod="openshift-marketplace/certified-operators-tpxl4" Nov 23 07:19:15 crc kubenswrapper[4559]: I1123 07:19:15.671021 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tpxl4" Nov 23 07:19:16 crc kubenswrapper[4559]: I1123 07:19:16.104504 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tpxl4"] Nov 23 07:19:16 crc kubenswrapper[4559]: I1123 07:19:16.759387 4559 generic.go:334] "Generic (PLEG): container finished" podID="b08ab74f-9543-4016-8d4d-00134d2bfe5a" containerID="702b2470f73379326eeefbc513558f24da90ba2f9419324206161ebd682c3149" exitCode=0 Nov 23 07:19:16 crc kubenswrapper[4559]: I1123 07:19:16.759478 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tpxl4" event={"ID":"b08ab74f-9543-4016-8d4d-00134d2bfe5a","Type":"ContainerDied","Data":"702b2470f73379326eeefbc513558f24da90ba2f9419324206161ebd682c3149"} Nov 23 07:19:16 crc kubenswrapper[4559]: I1123 07:19:16.759724 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tpxl4" event={"ID":"b08ab74f-9543-4016-8d4d-00134d2bfe5a","Type":"ContainerStarted","Data":"df60d071eceed8f8bb65c02fe9dff37def0fb2f96a4ab2ba676329c0137e2428"} Nov 23 07:19:17 crc kubenswrapper[4559]: I1123 07:19:17.767367 4559 generic.go:334] "Generic (PLEG): container finished" podID="14f31bb2-9f00-4931-a84a-900401189cc0" containerID="271234e32a4d3ae4d5cb127190f0ba0b823982ecfc2dddadd31f27d098dea80e" exitCode=0 Nov 23 07:19:17 crc kubenswrapper[4559]: I1123 07:19:17.767457 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" event={"ID":"14f31bb2-9f00-4931-a84a-900401189cc0","Type":"ContainerDied","Data":"271234e32a4d3ae4d5cb127190f0ba0b823982ecfc2dddadd31f27d098dea80e"} Nov 23 07:19:17 crc kubenswrapper[4559]: I1123 07:19:17.769408 4559 generic.go:334] "Generic (PLEG): container finished" podID="b08ab74f-9543-4016-8d4d-00134d2bfe5a" containerID="df63c8b4130e3b6d21f7bafa66672e1d089312fb9255aa7654a5d4173d93f4b0" exitCode=0 Nov 23 07:19:17 crc kubenswrapper[4559]: I1123 07:19:17.769437 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tpxl4" event={"ID":"b08ab74f-9543-4016-8d4d-00134d2bfe5a","Type":"ContainerDied","Data":"df63c8b4130e3b6d21f7bafa66672e1d089312fb9255aa7654a5d4173d93f4b0"} Nov 23 07:19:18 crc kubenswrapper[4559]: I1123 07:19:18.778818 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tpxl4" event={"ID":"b08ab74f-9543-4016-8d4d-00134d2bfe5a","Type":"ContainerStarted","Data":"dccf7f7f50bad2d8375a803abdcaff038bce2dd6bd9e617e11360a415a41a8f6"} Nov 23 07:19:18 crc kubenswrapper[4559]: I1123 07:19:18.825823 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tpxl4" podStartSLOduration=2.371697137 podStartE2EDuration="3.825802862s" podCreationTimestamp="2025-11-23 07:19:15 +0000 UTC" firstStartedPulling="2025-11-23 07:19:16.76160943 +0000 UTC m=+2058.783595044" lastFinishedPulling="2025-11-23 07:19:18.215715156 +0000 UTC m=+2060.237700769" observedRunningTime="2025-11-23 07:19:18.800005518 +0000 UTC m=+2060.821991132" watchObservedRunningTime="2025-11-23 07:19:18.825802862 +0000 UTC m=+2060.847788476" Nov 23 07:19:19 crc kubenswrapper[4559]: I1123 07:19:19.113864 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" Nov 23 07:19:19 crc kubenswrapper[4559]: I1123 07:19:19.244767 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-ssh-key\") pod \"14f31bb2-9f00-4931-a84a-900401189cc0\" (UID: \"14f31bb2-9f00-4931-a84a-900401189cc0\") " Nov 23 07:19:19 crc kubenswrapper[4559]: I1123 07:19:19.245190 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-telemetry-combined-ca-bundle\") pod \"14f31bb2-9f00-4931-a84a-900401189cc0\" (UID: \"14f31bb2-9f00-4931-a84a-900401189cc0\") " Nov 23 07:19:19 crc kubenswrapper[4559]: I1123 07:19:19.245549 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xxcz\" (UniqueName: \"kubernetes.io/projected/14f31bb2-9f00-4931-a84a-900401189cc0-kube-api-access-2xxcz\") pod \"14f31bb2-9f00-4931-a84a-900401189cc0\" (UID: \"14f31bb2-9f00-4931-a84a-900401189cc0\") " Nov 23 07:19:19 crc kubenswrapper[4559]: I1123 07:19:19.245601 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-ceilometer-compute-config-data-1\") pod \"14f31bb2-9f00-4931-a84a-900401189cc0\" (UID: \"14f31bb2-9f00-4931-a84a-900401189cc0\") " Nov 23 07:19:19 crc kubenswrapper[4559]: I1123 07:19:19.245720 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-ceilometer-compute-config-data-0\") pod \"14f31bb2-9f00-4931-a84a-900401189cc0\" (UID: \"14f31bb2-9f00-4931-a84a-900401189cc0\") " Nov 23 07:19:19 crc kubenswrapper[4559]: I1123 07:19:19.245830 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-ceilometer-compute-config-data-2\") pod \"14f31bb2-9f00-4931-a84a-900401189cc0\" (UID: \"14f31bb2-9f00-4931-a84a-900401189cc0\") " Nov 23 07:19:19 crc kubenswrapper[4559]: I1123 07:19:19.245931 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-inventory\") pod \"14f31bb2-9f00-4931-a84a-900401189cc0\" (UID: \"14f31bb2-9f00-4931-a84a-900401189cc0\") " Nov 23 07:19:19 crc kubenswrapper[4559]: I1123 07:19:19.252027 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14f31bb2-9f00-4931-a84a-900401189cc0-kube-api-access-2xxcz" (OuterVolumeSpecName: "kube-api-access-2xxcz") pod "14f31bb2-9f00-4931-a84a-900401189cc0" (UID: "14f31bb2-9f00-4931-a84a-900401189cc0"). InnerVolumeSpecName "kube-api-access-2xxcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:19:19 crc kubenswrapper[4559]: I1123 07:19:19.252818 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "14f31bb2-9f00-4931-a84a-900401189cc0" (UID: "14f31bb2-9f00-4931-a84a-900401189cc0"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:19:19 crc kubenswrapper[4559]: I1123 07:19:19.270070 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "14f31bb2-9f00-4931-a84a-900401189cc0" (UID: "14f31bb2-9f00-4931-a84a-900401189cc0"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:19:19 crc kubenswrapper[4559]: I1123 07:19:19.271120 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-inventory" (OuterVolumeSpecName: "inventory") pod "14f31bb2-9f00-4931-a84a-900401189cc0" (UID: "14f31bb2-9f00-4931-a84a-900401189cc0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:19:19 crc kubenswrapper[4559]: I1123 07:19:19.272211 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "14f31bb2-9f00-4931-a84a-900401189cc0" (UID: "14f31bb2-9f00-4931-a84a-900401189cc0"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:19:19 crc kubenswrapper[4559]: I1123 07:19:19.272500 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "14f31bb2-9f00-4931-a84a-900401189cc0" (UID: "14f31bb2-9f00-4931-a84a-900401189cc0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:19:19 crc kubenswrapper[4559]: I1123 07:19:19.272561 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "14f31bb2-9f00-4931-a84a-900401189cc0" (UID: "14f31bb2-9f00-4931-a84a-900401189cc0"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:19:19 crc kubenswrapper[4559]: I1123 07:19:19.351013 4559 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 23 07:19:19 crc kubenswrapper[4559]: I1123 07:19:19.351153 4559 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 07:19:19 crc kubenswrapper[4559]: I1123 07:19:19.351232 4559 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 07:19:19 crc kubenswrapper[4559]: I1123 07:19:19.351295 4559 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 07:19:19 crc kubenswrapper[4559]: I1123 07:19:19.351354 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xxcz\" (UniqueName: \"kubernetes.io/projected/14f31bb2-9f00-4931-a84a-900401189cc0-kube-api-access-2xxcz\") on node \"crc\" DevicePath \"\"" Nov 23 07:19:19 crc kubenswrapper[4559]: I1123 07:19:19.351408 4559 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 23 07:19:19 crc kubenswrapper[4559]: I1123 07:19:19.351458 4559 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/14f31bb2-9f00-4931-a84a-900401189cc0-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 23 07:19:19 crc kubenswrapper[4559]: I1123 07:19:19.786799 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" Nov 23 07:19:19 crc kubenswrapper[4559]: I1123 07:19:19.786781 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8" event={"ID":"14f31bb2-9f00-4931-a84a-900401189cc0","Type":"ContainerDied","Data":"aa7dcc04a8a566fd9fe384e790072cadb04fe9cda9f544ab2b133a36826fb33e"} Nov 23 07:19:19 crc kubenswrapper[4559]: I1123 07:19:19.787689 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa7dcc04a8a566fd9fe384e790072cadb04fe9cda9f544ab2b133a36826fb33e" Nov 23 07:19:25 crc kubenswrapper[4559]: I1123 07:19:25.672133 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tpxl4" Nov 23 07:19:25 crc kubenswrapper[4559]: I1123 07:19:25.672552 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tpxl4" Nov 23 07:19:25 crc kubenswrapper[4559]: I1123 07:19:25.703872 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tpxl4" Nov 23 07:19:25 crc kubenswrapper[4559]: I1123 07:19:25.863929 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tpxl4" Nov 23 07:19:25 crc kubenswrapper[4559]: I1123 07:19:25.929242 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tpxl4"] Nov 23 07:19:27 crc kubenswrapper[4559]: I1123 07:19:27.843802 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tpxl4" podUID="b08ab74f-9543-4016-8d4d-00134d2bfe5a" containerName="registry-server" containerID="cri-o://dccf7f7f50bad2d8375a803abdcaff038bce2dd6bd9e617e11360a415a41a8f6" gracePeriod=2 Nov 23 07:19:28 crc kubenswrapper[4559]: I1123 07:19:28.223611 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tpxl4" Nov 23 07:19:28 crc kubenswrapper[4559]: I1123 07:19:28.396808 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpjkj\" (UniqueName: \"kubernetes.io/projected/b08ab74f-9543-4016-8d4d-00134d2bfe5a-kube-api-access-wpjkj\") pod \"b08ab74f-9543-4016-8d4d-00134d2bfe5a\" (UID: \"b08ab74f-9543-4016-8d4d-00134d2bfe5a\") " Nov 23 07:19:28 crc kubenswrapper[4559]: I1123 07:19:28.396906 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b08ab74f-9543-4016-8d4d-00134d2bfe5a-utilities\") pod \"b08ab74f-9543-4016-8d4d-00134d2bfe5a\" (UID: \"b08ab74f-9543-4016-8d4d-00134d2bfe5a\") " Nov 23 07:19:28 crc kubenswrapper[4559]: I1123 07:19:28.397081 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b08ab74f-9543-4016-8d4d-00134d2bfe5a-catalog-content\") pod \"b08ab74f-9543-4016-8d4d-00134d2bfe5a\" (UID: \"b08ab74f-9543-4016-8d4d-00134d2bfe5a\") " Nov 23 07:19:28 crc kubenswrapper[4559]: I1123 07:19:28.397567 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b08ab74f-9543-4016-8d4d-00134d2bfe5a-utilities" (OuterVolumeSpecName: "utilities") pod "b08ab74f-9543-4016-8d4d-00134d2bfe5a" (UID: "b08ab74f-9543-4016-8d4d-00134d2bfe5a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:19:28 crc kubenswrapper[4559]: I1123 07:19:28.397854 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b08ab74f-9543-4016-8d4d-00134d2bfe5a-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:19:28 crc kubenswrapper[4559]: I1123 07:19:28.401963 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b08ab74f-9543-4016-8d4d-00134d2bfe5a-kube-api-access-wpjkj" (OuterVolumeSpecName: "kube-api-access-wpjkj") pod "b08ab74f-9543-4016-8d4d-00134d2bfe5a" (UID: "b08ab74f-9543-4016-8d4d-00134d2bfe5a"). InnerVolumeSpecName "kube-api-access-wpjkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:19:28 crc kubenswrapper[4559]: I1123 07:19:28.433284 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b08ab74f-9543-4016-8d4d-00134d2bfe5a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b08ab74f-9543-4016-8d4d-00134d2bfe5a" (UID: "b08ab74f-9543-4016-8d4d-00134d2bfe5a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:19:28 crc kubenswrapper[4559]: I1123 07:19:28.500122 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b08ab74f-9543-4016-8d4d-00134d2bfe5a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:19:28 crc kubenswrapper[4559]: I1123 07:19:28.500156 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpjkj\" (UniqueName: \"kubernetes.io/projected/b08ab74f-9543-4016-8d4d-00134d2bfe5a-kube-api-access-wpjkj\") on node \"crc\" DevicePath \"\"" Nov 23 07:19:28 crc kubenswrapper[4559]: I1123 07:19:28.854958 4559 generic.go:334] "Generic (PLEG): container finished" podID="b08ab74f-9543-4016-8d4d-00134d2bfe5a" containerID="dccf7f7f50bad2d8375a803abdcaff038bce2dd6bd9e617e11360a415a41a8f6" exitCode=0 Nov 23 07:19:28 crc kubenswrapper[4559]: I1123 07:19:28.855029 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tpxl4" Nov 23 07:19:28 crc kubenswrapper[4559]: I1123 07:19:28.855048 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tpxl4" event={"ID":"b08ab74f-9543-4016-8d4d-00134d2bfe5a","Type":"ContainerDied","Data":"dccf7f7f50bad2d8375a803abdcaff038bce2dd6bd9e617e11360a415a41a8f6"} Nov 23 07:19:28 crc kubenswrapper[4559]: I1123 07:19:28.855391 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tpxl4" event={"ID":"b08ab74f-9543-4016-8d4d-00134d2bfe5a","Type":"ContainerDied","Data":"df60d071eceed8f8bb65c02fe9dff37def0fb2f96a4ab2ba676329c0137e2428"} Nov 23 07:19:28 crc kubenswrapper[4559]: I1123 07:19:28.855415 4559 scope.go:117] "RemoveContainer" containerID="dccf7f7f50bad2d8375a803abdcaff038bce2dd6bd9e617e11360a415a41a8f6" Nov 23 07:19:28 crc kubenswrapper[4559]: I1123 07:19:28.875767 4559 scope.go:117] "RemoveContainer" containerID="df63c8b4130e3b6d21f7bafa66672e1d089312fb9255aa7654a5d4173d93f4b0" Nov 23 07:19:28 crc kubenswrapper[4559]: I1123 07:19:28.880806 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tpxl4"] Nov 23 07:19:28 crc kubenswrapper[4559]: I1123 07:19:28.887188 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tpxl4"] Nov 23 07:19:28 crc kubenswrapper[4559]: I1123 07:19:28.901985 4559 scope.go:117] "RemoveContainer" containerID="702b2470f73379326eeefbc513558f24da90ba2f9419324206161ebd682c3149" Nov 23 07:19:28 crc kubenswrapper[4559]: I1123 07:19:28.923779 4559 scope.go:117] "RemoveContainer" containerID="dccf7f7f50bad2d8375a803abdcaff038bce2dd6bd9e617e11360a415a41a8f6" Nov 23 07:19:28 crc kubenswrapper[4559]: E1123 07:19:28.924127 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dccf7f7f50bad2d8375a803abdcaff038bce2dd6bd9e617e11360a415a41a8f6\": container with ID starting with dccf7f7f50bad2d8375a803abdcaff038bce2dd6bd9e617e11360a415a41a8f6 not found: ID does not exist" containerID="dccf7f7f50bad2d8375a803abdcaff038bce2dd6bd9e617e11360a415a41a8f6" Nov 23 07:19:28 crc kubenswrapper[4559]: I1123 07:19:28.924155 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dccf7f7f50bad2d8375a803abdcaff038bce2dd6bd9e617e11360a415a41a8f6"} err="failed to get container status \"dccf7f7f50bad2d8375a803abdcaff038bce2dd6bd9e617e11360a415a41a8f6\": rpc error: code = NotFound desc = could not find container \"dccf7f7f50bad2d8375a803abdcaff038bce2dd6bd9e617e11360a415a41a8f6\": container with ID starting with dccf7f7f50bad2d8375a803abdcaff038bce2dd6bd9e617e11360a415a41a8f6 not found: ID does not exist" Nov 23 07:19:28 crc kubenswrapper[4559]: I1123 07:19:28.924176 4559 scope.go:117] "RemoveContainer" containerID="df63c8b4130e3b6d21f7bafa66672e1d089312fb9255aa7654a5d4173d93f4b0" Nov 23 07:19:28 crc kubenswrapper[4559]: E1123 07:19:28.924537 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df63c8b4130e3b6d21f7bafa66672e1d089312fb9255aa7654a5d4173d93f4b0\": container with ID starting with df63c8b4130e3b6d21f7bafa66672e1d089312fb9255aa7654a5d4173d93f4b0 not found: ID does not exist" containerID="df63c8b4130e3b6d21f7bafa66672e1d089312fb9255aa7654a5d4173d93f4b0" Nov 23 07:19:28 crc kubenswrapper[4559]: I1123 07:19:28.924577 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df63c8b4130e3b6d21f7bafa66672e1d089312fb9255aa7654a5d4173d93f4b0"} err="failed to get container status \"df63c8b4130e3b6d21f7bafa66672e1d089312fb9255aa7654a5d4173d93f4b0\": rpc error: code = NotFound desc = could not find container \"df63c8b4130e3b6d21f7bafa66672e1d089312fb9255aa7654a5d4173d93f4b0\": container with ID starting with df63c8b4130e3b6d21f7bafa66672e1d089312fb9255aa7654a5d4173d93f4b0 not found: ID does not exist" Nov 23 07:19:28 crc kubenswrapper[4559]: I1123 07:19:28.924627 4559 scope.go:117] "RemoveContainer" containerID="702b2470f73379326eeefbc513558f24da90ba2f9419324206161ebd682c3149" Nov 23 07:19:28 crc kubenswrapper[4559]: E1123 07:19:28.925033 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"702b2470f73379326eeefbc513558f24da90ba2f9419324206161ebd682c3149\": container with ID starting with 702b2470f73379326eeefbc513558f24da90ba2f9419324206161ebd682c3149 not found: ID does not exist" containerID="702b2470f73379326eeefbc513558f24da90ba2f9419324206161ebd682c3149" Nov 23 07:19:28 crc kubenswrapper[4559]: I1123 07:19:28.925053 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"702b2470f73379326eeefbc513558f24da90ba2f9419324206161ebd682c3149"} err="failed to get container status \"702b2470f73379326eeefbc513558f24da90ba2f9419324206161ebd682c3149\": rpc error: code = NotFound desc = could not find container \"702b2470f73379326eeefbc513558f24da90ba2f9419324206161ebd682c3149\": container with ID starting with 702b2470f73379326eeefbc513558f24da90ba2f9419324206161ebd682c3149 not found: ID does not exist" Nov 23 07:19:30 crc kubenswrapper[4559]: I1123 07:19:30.281947 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b08ab74f-9543-4016-8d4d-00134d2bfe5a" path="/var/lib/kubelet/pods/b08ab74f-9543-4016-8d4d-00134d2bfe5a/volumes" Nov 23 07:19:56 crc kubenswrapper[4559]: I1123 07:19:56.167250 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:19:56 crc kubenswrapper[4559]: I1123 07:19:56.167681 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.530461 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Nov 23 07:20:05 crc kubenswrapper[4559]: E1123 07:20:05.531226 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b08ab74f-9543-4016-8d4d-00134d2bfe5a" containerName="extract-content" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.531241 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="b08ab74f-9543-4016-8d4d-00134d2bfe5a" containerName="extract-content" Nov 23 07:20:05 crc kubenswrapper[4559]: E1123 07:20:05.531252 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f31bb2-9f00-4931-a84a-900401189cc0" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.531258 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f31bb2-9f00-4931-a84a-900401189cc0" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 23 07:20:05 crc kubenswrapper[4559]: E1123 07:20:05.531270 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b08ab74f-9543-4016-8d4d-00134d2bfe5a" containerName="extract-utilities" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.531276 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="b08ab74f-9543-4016-8d4d-00134d2bfe5a" containerName="extract-utilities" Nov 23 07:20:05 crc kubenswrapper[4559]: E1123 07:20:05.531294 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b08ab74f-9543-4016-8d4d-00134d2bfe5a" containerName="registry-server" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.531299 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="b08ab74f-9543-4016-8d4d-00134d2bfe5a" containerName="registry-server" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.531499 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="b08ab74f-9543-4016-8d4d-00134d2bfe5a" containerName="registry-server" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.531510 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f31bb2-9f00-4931-a84a-900401189cc0" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.532909 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.534342 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.534619 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.534750 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.534802 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-pdp84" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.537775 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.685346 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fb299f6e-9a30-4d29-95db-a0cbceb573ea-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.685584 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb299f6e-9a30-4d29-95db-a0cbceb573ea-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.685703 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47ngx\" (UniqueName: \"kubernetes.io/projected/fb299f6e-9a30-4d29-95db-a0cbceb573ea-kube-api-access-47ngx\") pod \"tempest-tests-tempest\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.685816 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fb299f6e-9a30-4d29-95db-a0cbceb573ea-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.685940 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/fb299f6e-9a30-4d29-95db-a0cbceb573ea-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.686061 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/fb299f6e-9a30-4d29-95db-a0cbceb573ea-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.686157 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.686242 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb299f6e-9a30-4d29-95db-a0cbceb573ea-config-data\") pod \"tempest-tests-tempest\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.686304 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/fb299f6e-9a30-4d29-95db-a0cbceb573ea-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.788479 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/fb299f6e-9a30-4d29-95db-a0cbceb573ea-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.788590 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/fb299f6e-9a30-4d29-95db-a0cbceb573ea-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.788615 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.788682 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb299f6e-9a30-4d29-95db-a0cbceb573ea-config-data\") pod \"tempest-tests-tempest\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.788697 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/fb299f6e-9a30-4d29-95db-a0cbceb573ea-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.788736 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fb299f6e-9a30-4d29-95db-a0cbceb573ea-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.788756 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb299f6e-9a30-4d29-95db-a0cbceb573ea-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.788785 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47ngx\" (UniqueName: \"kubernetes.io/projected/fb299f6e-9a30-4d29-95db-a0cbceb573ea-kube-api-access-47ngx\") pod \"tempest-tests-tempest\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.788842 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fb299f6e-9a30-4d29-95db-a0cbceb573ea-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.788895 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/fb299f6e-9a30-4d29-95db-a0cbceb573ea-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.789082 4559 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.789690 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fb299f6e-9a30-4d29-95db-a0cbceb573ea-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.789917 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/fb299f6e-9a30-4d29-95db-a0cbceb573ea-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.790272 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb299f6e-9a30-4d29-95db-a0cbceb573ea-config-data\") pod \"tempest-tests-tempest\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.793753 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fb299f6e-9a30-4d29-95db-a0cbceb573ea-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.793911 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/fb299f6e-9a30-4d29-95db-a0cbceb573ea-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.794266 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb299f6e-9a30-4d29-95db-a0cbceb573ea-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.801178 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47ngx\" (UniqueName: \"kubernetes.io/projected/fb299f6e-9a30-4d29-95db-a0cbceb573ea-kube-api-access-47ngx\") pod \"tempest-tests-tempest\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.809415 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " pod="openstack/tempest-tests-tempest" Nov 23 07:20:05 crc kubenswrapper[4559]: I1123 07:20:05.847176 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 23 07:20:06 crc kubenswrapper[4559]: I1123 07:20:06.199159 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 23 07:20:07 crc kubenswrapper[4559]: I1123 07:20:07.095388 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"fb299f6e-9a30-4d29-95db-a0cbceb573ea","Type":"ContainerStarted","Data":"620011a02e5d17c29b5afd433ca1db95a9b42700270ae3b6a76ea735f6879236"} Nov 23 07:20:26 crc kubenswrapper[4559]: I1123 07:20:26.166551 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:20:26 crc kubenswrapper[4559]: I1123 07:20:26.166952 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:20:40 crc kubenswrapper[4559]: I1123 07:20:40.836078 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-29944"] Nov 23 07:20:40 crc kubenswrapper[4559]: I1123 07:20:40.838816 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-29944" Nov 23 07:20:40 crc kubenswrapper[4559]: I1123 07:20:40.845682 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-29944"] Nov 23 07:20:40 crc kubenswrapper[4559]: I1123 07:20:40.931258 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9-utilities\") pod \"community-operators-29944\" (UID: \"9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9\") " pod="openshift-marketplace/community-operators-29944" Nov 23 07:20:40 crc kubenswrapper[4559]: I1123 07:20:40.931817 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9-catalog-content\") pod \"community-operators-29944\" (UID: \"9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9\") " pod="openshift-marketplace/community-operators-29944" Nov 23 07:20:40 crc kubenswrapper[4559]: I1123 07:20:40.932191 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zrnl\" (UniqueName: \"kubernetes.io/projected/9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9-kube-api-access-9zrnl\") pod \"community-operators-29944\" (UID: \"9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9\") " pod="openshift-marketplace/community-operators-29944" Nov 23 07:20:41 crc kubenswrapper[4559]: I1123 07:20:41.035190 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9-utilities\") pod \"community-operators-29944\" (UID: \"9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9\") " pod="openshift-marketplace/community-operators-29944" Nov 23 07:20:41 crc kubenswrapper[4559]: I1123 07:20:41.035282 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9-catalog-content\") pod \"community-operators-29944\" (UID: \"9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9\") " pod="openshift-marketplace/community-operators-29944" Nov 23 07:20:41 crc kubenswrapper[4559]: I1123 07:20:41.035359 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zrnl\" (UniqueName: \"kubernetes.io/projected/9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9-kube-api-access-9zrnl\") pod \"community-operators-29944\" (UID: \"9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9\") " pod="openshift-marketplace/community-operators-29944" Nov 23 07:20:41 crc kubenswrapper[4559]: I1123 07:20:41.036180 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9-utilities\") pod \"community-operators-29944\" (UID: \"9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9\") " pod="openshift-marketplace/community-operators-29944" Nov 23 07:20:41 crc kubenswrapper[4559]: I1123 07:20:41.036317 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9-catalog-content\") pod \"community-operators-29944\" (UID: \"9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9\") " pod="openshift-marketplace/community-operators-29944" Nov 23 07:20:41 crc kubenswrapper[4559]: I1123 07:20:41.070525 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zrnl\" (UniqueName: \"kubernetes.io/projected/9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9-kube-api-access-9zrnl\") pod \"community-operators-29944\" (UID: \"9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9\") " pod="openshift-marketplace/community-operators-29944" Nov 23 07:20:41 crc kubenswrapper[4559]: I1123 07:20:41.169338 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-29944" Nov 23 07:20:45 crc kubenswrapper[4559]: E1123 07:20:45.944852 4559 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:8e43c662a6abf8c9a07ada252f8dc6af" Nov 23 07:20:45 crc kubenswrapper[4559]: E1123 07:20:45.945153 4559 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:8e43c662a6abf8c9a07ada252f8dc6af" Nov 23 07:20:45 crc kubenswrapper[4559]: E1123 07:20:45.945272 4559 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:8e43c662a6abf8c9a07ada252f8dc6af,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-47ngx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(fb299f6e-9a30-4d29-95db-a0cbceb573ea): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 07:20:45 crc kubenswrapper[4559]: E1123 07:20:45.947341 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="fb299f6e-9a30-4d29-95db-a0cbceb573ea" Nov 23 07:20:46 crc kubenswrapper[4559]: I1123 07:20:46.301508 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-29944"] Nov 23 07:20:46 crc kubenswrapper[4559]: I1123 07:20:46.375270 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-29944" event={"ID":"9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9","Type":"ContainerStarted","Data":"4c4eea7b23c7ed904238af8326dbfcdb53399d0a41183f4cb6b458e2c65540cc"} Nov 23 07:20:46 crc kubenswrapper[4559]: E1123 07:20:46.376699 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:8e43c662a6abf8c9a07ada252f8dc6af\\\"\"" pod="openstack/tempest-tests-tempest" podUID="fb299f6e-9a30-4d29-95db-a0cbceb573ea" Nov 23 07:20:47 crc kubenswrapper[4559]: I1123 07:20:47.383444 4559 generic.go:334] "Generic (PLEG): container finished" podID="9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9" containerID="23ca76dad090d8d08b82ad35b4505a5db9339d8171710c05411bfe32b419be8c" exitCode=0 Nov 23 07:20:47 crc kubenswrapper[4559]: I1123 07:20:47.383484 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-29944" event={"ID":"9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9","Type":"ContainerDied","Data":"23ca76dad090d8d08b82ad35b4505a5db9339d8171710c05411bfe32b419be8c"} Nov 23 07:20:47 crc kubenswrapper[4559]: I1123 07:20:47.385166 4559 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 07:20:48 crc kubenswrapper[4559]: I1123 07:20:48.391025 4559 generic.go:334] "Generic (PLEG): container finished" podID="9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9" containerID="acc2f2bbe52c81c3489afa0f7f57c557281e27aea92bbcf1e264a999e92bedc8" exitCode=0 Nov 23 07:20:48 crc kubenswrapper[4559]: I1123 07:20:48.391103 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-29944" event={"ID":"9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9","Type":"ContainerDied","Data":"acc2f2bbe52c81c3489afa0f7f57c557281e27aea92bbcf1e264a999e92bedc8"} Nov 23 07:20:49 crc kubenswrapper[4559]: I1123 07:20:49.400859 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-29944" event={"ID":"9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9","Type":"ContainerStarted","Data":"438d314a44fba4dae633ee619c49889ee8140c4be56918d9d47b8b9617a2f39d"} Nov 23 07:20:49 crc kubenswrapper[4559]: I1123 07:20:49.412792 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-29944" podStartSLOduration=7.904943011 podStartE2EDuration="9.412777288s" podCreationTimestamp="2025-11-23 07:20:40 +0000 UTC" firstStartedPulling="2025-11-23 07:20:47.384954377 +0000 UTC m=+2149.406939992" lastFinishedPulling="2025-11-23 07:20:48.892788654 +0000 UTC m=+2150.914774269" observedRunningTime="2025-11-23 07:20:49.411941607 +0000 UTC m=+2151.433927221" watchObservedRunningTime="2025-11-23 07:20:49.412777288 +0000 UTC m=+2151.434762902" Nov 23 07:20:51 crc kubenswrapper[4559]: I1123 07:20:51.170882 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-29944" Nov 23 07:20:51 crc kubenswrapper[4559]: I1123 07:20:51.171022 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-29944" Nov 23 07:20:51 crc kubenswrapper[4559]: I1123 07:20:51.202464 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-29944" Nov 23 07:20:56 crc kubenswrapper[4559]: I1123 07:20:56.166945 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:20:56 crc kubenswrapper[4559]: I1123 07:20:56.167290 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:20:56 crc kubenswrapper[4559]: I1123 07:20:56.167335 4559 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" Nov 23 07:20:56 crc kubenswrapper[4559]: I1123 07:20:56.167852 4559 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f"} pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 07:20:56 crc kubenswrapper[4559]: I1123 07:20:56.167906 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" containerID="cri-o://e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f" gracePeriod=600 Nov 23 07:20:56 crc kubenswrapper[4559]: E1123 07:20:56.284536 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:20:56 crc kubenswrapper[4559]: I1123 07:20:56.447376 4559 generic.go:334] "Generic (PLEG): container finished" podID="4731beee-0cac-4189-8a70-743b0b709095" containerID="e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f" exitCode=0 Nov 23 07:20:56 crc kubenswrapper[4559]: I1123 07:20:56.447444 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerDied","Data":"e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f"} Nov 23 07:20:56 crc kubenswrapper[4559]: I1123 07:20:56.447766 4559 scope.go:117] "RemoveContainer" containerID="eb8bcd23b2c4b87ac90b1547ee493c8aff582ddd59e852cc40a1f99ebbec9b19" Nov 23 07:20:56 crc kubenswrapper[4559]: I1123 07:20:56.448284 4559 scope.go:117] "RemoveContainer" containerID="e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f" Nov 23 07:20:56 crc kubenswrapper[4559]: E1123 07:20:56.448554 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:20:59 crc kubenswrapper[4559]: I1123 07:20:59.468788 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"fb299f6e-9a30-4d29-95db-a0cbceb573ea","Type":"ContainerStarted","Data":"3fb16f9f3381f4ad7ea54d90253d1175d802011895e8e41536ac004dae8676da"} Nov 23 07:20:59 crc kubenswrapper[4559]: I1123 07:20:59.481746 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.772206626 podStartE2EDuration="55.48173196s" podCreationTimestamp="2025-11-23 07:20:04 +0000 UTC" firstStartedPulling="2025-11-23 07:20:06.202743695 +0000 UTC m=+2108.224729309" lastFinishedPulling="2025-11-23 07:20:57.912269029 +0000 UTC m=+2159.934254643" observedRunningTime="2025-11-23 07:20:59.481241979 +0000 UTC m=+2161.503227594" watchObservedRunningTime="2025-11-23 07:20:59.48173196 +0000 UTC m=+2161.503717574" Nov 23 07:21:01 crc kubenswrapper[4559]: I1123 07:21:01.214676 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-29944" Nov 23 07:21:01 crc kubenswrapper[4559]: I1123 07:21:01.247842 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-29944"] Nov 23 07:21:01 crc kubenswrapper[4559]: I1123 07:21:01.481022 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-29944" podUID="9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9" containerName="registry-server" containerID="cri-o://438d314a44fba4dae633ee619c49889ee8140c4be56918d9d47b8b9617a2f39d" gracePeriod=2 Nov 23 07:21:01 crc kubenswrapper[4559]: I1123 07:21:01.846600 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-29944" Nov 23 07:21:02 crc kubenswrapper[4559]: I1123 07:21:02.018574 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zrnl\" (UniqueName: \"kubernetes.io/projected/9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9-kube-api-access-9zrnl\") pod \"9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9\" (UID: \"9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9\") " Nov 23 07:21:02 crc kubenswrapper[4559]: I1123 07:21:02.018866 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9-utilities\") pod \"9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9\" (UID: \"9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9\") " Nov 23 07:21:02 crc kubenswrapper[4559]: I1123 07:21:02.018891 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9-catalog-content\") pod \"9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9\" (UID: \"9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9\") " Nov 23 07:21:02 crc kubenswrapper[4559]: I1123 07:21:02.019635 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9-utilities" (OuterVolumeSpecName: "utilities") pod "9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9" (UID: "9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:21:02 crc kubenswrapper[4559]: I1123 07:21:02.032788 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9-kube-api-access-9zrnl" (OuterVolumeSpecName: "kube-api-access-9zrnl") pod "9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9" (UID: "9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9"). InnerVolumeSpecName "kube-api-access-9zrnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:21:02 crc kubenswrapper[4559]: I1123 07:21:02.056725 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9" (UID: "9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:21:02 crc kubenswrapper[4559]: I1123 07:21:02.120878 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zrnl\" (UniqueName: \"kubernetes.io/projected/9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9-kube-api-access-9zrnl\") on node \"crc\" DevicePath \"\"" Nov 23 07:21:02 crc kubenswrapper[4559]: I1123 07:21:02.120904 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:21:02 crc kubenswrapper[4559]: I1123 07:21:02.120917 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:21:02 crc kubenswrapper[4559]: I1123 07:21:02.490957 4559 generic.go:334] "Generic (PLEG): container finished" podID="9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9" containerID="438d314a44fba4dae633ee619c49889ee8140c4be56918d9d47b8b9617a2f39d" exitCode=0 Nov 23 07:21:02 crc kubenswrapper[4559]: I1123 07:21:02.490996 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-29944" event={"ID":"9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9","Type":"ContainerDied","Data":"438d314a44fba4dae633ee619c49889ee8140c4be56918d9d47b8b9617a2f39d"} Nov 23 07:21:02 crc kubenswrapper[4559]: I1123 07:21:02.491025 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-29944" event={"ID":"9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9","Type":"ContainerDied","Data":"4c4eea7b23c7ed904238af8326dbfcdb53399d0a41183f4cb6b458e2c65540cc"} Nov 23 07:21:02 crc kubenswrapper[4559]: I1123 07:21:02.491036 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-29944" Nov 23 07:21:02 crc kubenswrapper[4559]: I1123 07:21:02.491041 4559 scope.go:117] "RemoveContainer" containerID="438d314a44fba4dae633ee619c49889ee8140c4be56918d9d47b8b9617a2f39d" Nov 23 07:21:02 crc kubenswrapper[4559]: I1123 07:21:02.507476 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-29944"] Nov 23 07:21:02 crc kubenswrapper[4559]: I1123 07:21:02.508338 4559 scope.go:117] "RemoveContainer" containerID="acc2f2bbe52c81c3489afa0f7f57c557281e27aea92bbcf1e264a999e92bedc8" Nov 23 07:21:02 crc kubenswrapper[4559]: I1123 07:21:02.513782 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-29944"] Nov 23 07:21:02 crc kubenswrapper[4559]: I1123 07:21:02.536577 4559 scope.go:117] "RemoveContainer" containerID="23ca76dad090d8d08b82ad35b4505a5db9339d8171710c05411bfe32b419be8c" Nov 23 07:21:02 crc kubenswrapper[4559]: I1123 07:21:02.561793 4559 scope.go:117] "RemoveContainer" containerID="438d314a44fba4dae633ee619c49889ee8140c4be56918d9d47b8b9617a2f39d" Nov 23 07:21:02 crc kubenswrapper[4559]: E1123 07:21:02.562059 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"438d314a44fba4dae633ee619c49889ee8140c4be56918d9d47b8b9617a2f39d\": container with ID starting with 438d314a44fba4dae633ee619c49889ee8140c4be56918d9d47b8b9617a2f39d not found: ID does not exist" containerID="438d314a44fba4dae633ee619c49889ee8140c4be56918d9d47b8b9617a2f39d" Nov 23 07:21:02 crc kubenswrapper[4559]: I1123 07:21:02.562086 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"438d314a44fba4dae633ee619c49889ee8140c4be56918d9d47b8b9617a2f39d"} err="failed to get container status \"438d314a44fba4dae633ee619c49889ee8140c4be56918d9d47b8b9617a2f39d\": rpc error: code = NotFound desc = could not find container \"438d314a44fba4dae633ee619c49889ee8140c4be56918d9d47b8b9617a2f39d\": container with ID starting with 438d314a44fba4dae633ee619c49889ee8140c4be56918d9d47b8b9617a2f39d not found: ID does not exist" Nov 23 07:21:02 crc kubenswrapper[4559]: I1123 07:21:02.562112 4559 scope.go:117] "RemoveContainer" containerID="acc2f2bbe52c81c3489afa0f7f57c557281e27aea92bbcf1e264a999e92bedc8" Nov 23 07:21:02 crc kubenswrapper[4559]: E1123 07:21:02.562507 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acc2f2bbe52c81c3489afa0f7f57c557281e27aea92bbcf1e264a999e92bedc8\": container with ID starting with acc2f2bbe52c81c3489afa0f7f57c557281e27aea92bbcf1e264a999e92bedc8 not found: ID does not exist" containerID="acc2f2bbe52c81c3489afa0f7f57c557281e27aea92bbcf1e264a999e92bedc8" Nov 23 07:21:02 crc kubenswrapper[4559]: I1123 07:21:02.562555 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acc2f2bbe52c81c3489afa0f7f57c557281e27aea92bbcf1e264a999e92bedc8"} err="failed to get container status \"acc2f2bbe52c81c3489afa0f7f57c557281e27aea92bbcf1e264a999e92bedc8\": rpc error: code = NotFound desc = could not find container \"acc2f2bbe52c81c3489afa0f7f57c557281e27aea92bbcf1e264a999e92bedc8\": container with ID starting with acc2f2bbe52c81c3489afa0f7f57c557281e27aea92bbcf1e264a999e92bedc8 not found: ID does not exist" Nov 23 07:21:02 crc kubenswrapper[4559]: I1123 07:21:02.562581 4559 scope.go:117] "RemoveContainer" containerID="23ca76dad090d8d08b82ad35b4505a5db9339d8171710c05411bfe32b419be8c" Nov 23 07:21:02 crc kubenswrapper[4559]: E1123 07:21:02.562836 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23ca76dad090d8d08b82ad35b4505a5db9339d8171710c05411bfe32b419be8c\": container with ID starting with 23ca76dad090d8d08b82ad35b4505a5db9339d8171710c05411bfe32b419be8c not found: ID does not exist" containerID="23ca76dad090d8d08b82ad35b4505a5db9339d8171710c05411bfe32b419be8c" Nov 23 07:21:02 crc kubenswrapper[4559]: I1123 07:21:02.562862 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23ca76dad090d8d08b82ad35b4505a5db9339d8171710c05411bfe32b419be8c"} err="failed to get container status \"23ca76dad090d8d08b82ad35b4505a5db9339d8171710c05411bfe32b419be8c\": rpc error: code = NotFound desc = could not find container \"23ca76dad090d8d08b82ad35b4505a5db9339d8171710c05411bfe32b419be8c\": container with ID starting with 23ca76dad090d8d08b82ad35b4505a5db9339d8171710c05411bfe32b419be8c not found: ID does not exist" Nov 23 07:21:04 crc kubenswrapper[4559]: I1123 07:21:04.281323 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9" path="/var/lib/kubelet/pods/9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9/volumes" Nov 23 07:21:09 crc kubenswrapper[4559]: I1123 07:21:09.273420 4559 scope.go:117] "RemoveContainer" containerID="e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f" Nov 23 07:21:09 crc kubenswrapper[4559]: E1123 07:21:09.274360 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:21:24 crc kubenswrapper[4559]: I1123 07:21:24.277684 4559 scope.go:117] "RemoveContainer" containerID="e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f" Nov 23 07:21:24 crc kubenswrapper[4559]: E1123 07:21:24.278266 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:21:37 crc kubenswrapper[4559]: I1123 07:21:37.273237 4559 scope.go:117] "RemoveContainer" containerID="e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f" Nov 23 07:21:37 crc kubenswrapper[4559]: E1123 07:21:37.273885 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:21:48 crc kubenswrapper[4559]: I1123 07:21:48.278981 4559 scope.go:117] "RemoveContainer" containerID="e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f" Nov 23 07:21:48 crc kubenswrapper[4559]: E1123 07:21:48.279511 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:22:02 crc kubenswrapper[4559]: I1123 07:22:02.273654 4559 scope.go:117] "RemoveContainer" containerID="e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f" Nov 23 07:22:02 crc kubenswrapper[4559]: E1123 07:22:02.274185 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:22:16 crc kubenswrapper[4559]: I1123 07:22:16.273380 4559 scope.go:117] "RemoveContainer" containerID="e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f" Nov 23 07:22:16 crc kubenswrapper[4559]: E1123 07:22:16.274015 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:22:29 crc kubenswrapper[4559]: I1123 07:22:29.273917 4559 scope.go:117] "RemoveContainer" containerID="e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f" Nov 23 07:22:29 crc kubenswrapper[4559]: E1123 07:22:29.274828 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:22:39 crc kubenswrapper[4559]: I1123 07:22:39.642057 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wj2km"] Nov 23 07:22:39 crc kubenswrapper[4559]: E1123 07:22:39.642758 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9" containerName="extract-utilities" Nov 23 07:22:39 crc kubenswrapper[4559]: I1123 07:22:39.642774 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9" containerName="extract-utilities" Nov 23 07:22:39 crc kubenswrapper[4559]: E1123 07:22:39.642799 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9" containerName="extract-content" Nov 23 07:22:39 crc kubenswrapper[4559]: I1123 07:22:39.642806 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9" containerName="extract-content" Nov 23 07:22:39 crc kubenswrapper[4559]: E1123 07:22:39.642814 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9" containerName="registry-server" Nov 23 07:22:39 crc kubenswrapper[4559]: I1123 07:22:39.642819 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9" containerName="registry-server" Nov 23 07:22:39 crc kubenswrapper[4559]: I1123 07:22:39.642987 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c0a916c-2f54-4021-ad81-9ad4a4d8c6e9" containerName="registry-server" Nov 23 07:22:39 crc kubenswrapper[4559]: I1123 07:22:39.644101 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wj2km" Nov 23 07:22:39 crc kubenswrapper[4559]: I1123 07:22:39.650268 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2acd71c9-2eb5-49c2-9e2f-310b05980b3d-utilities\") pod \"redhat-marketplace-wj2km\" (UID: \"2acd71c9-2eb5-49c2-9e2f-310b05980b3d\") " pod="openshift-marketplace/redhat-marketplace-wj2km" Nov 23 07:22:39 crc kubenswrapper[4559]: I1123 07:22:39.650526 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbp9w\" (UniqueName: \"kubernetes.io/projected/2acd71c9-2eb5-49c2-9e2f-310b05980b3d-kube-api-access-qbp9w\") pod \"redhat-marketplace-wj2km\" (UID: \"2acd71c9-2eb5-49c2-9e2f-310b05980b3d\") " pod="openshift-marketplace/redhat-marketplace-wj2km" Nov 23 07:22:39 crc kubenswrapper[4559]: I1123 07:22:39.650597 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2acd71c9-2eb5-49c2-9e2f-310b05980b3d-catalog-content\") pod \"redhat-marketplace-wj2km\" (UID: \"2acd71c9-2eb5-49c2-9e2f-310b05980b3d\") " pod="openshift-marketplace/redhat-marketplace-wj2km" Nov 23 07:22:39 crc kubenswrapper[4559]: I1123 07:22:39.652378 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wj2km"] Nov 23 07:22:39 crc kubenswrapper[4559]: I1123 07:22:39.752374 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2acd71c9-2eb5-49c2-9e2f-310b05980b3d-utilities\") pod \"redhat-marketplace-wj2km\" (UID: \"2acd71c9-2eb5-49c2-9e2f-310b05980b3d\") " pod="openshift-marketplace/redhat-marketplace-wj2km" Nov 23 07:22:39 crc kubenswrapper[4559]: I1123 07:22:39.752684 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbp9w\" (UniqueName: \"kubernetes.io/projected/2acd71c9-2eb5-49c2-9e2f-310b05980b3d-kube-api-access-qbp9w\") pod \"redhat-marketplace-wj2km\" (UID: \"2acd71c9-2eb5-49c2-9e2f-310b05980b3d\") " pod="openshift-marketplace/redhat-marketplace-wj2km" Nov 23 07:22:39 crc kubenswrapper[4559]: I1123 07:22:39.752736 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2acd71c9-2eb5-49c2-9e2f-310b05980b3d-catalog-content\") pod \"redhat-marketplace-wj2km\" (UID: \"2acd71c9-2eb5-49c2-9e2f-310b05980b3d\") " pod="openshift-marketplace/redhat-marketplace-wj2km" Nov 23 07:22:39 crc kubenswrapper[4559]: I1123 07:22:39.753001 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2acd71c9-2eb5-49c2-9e2f-310b05980b3d-utilities\") pod \"redhat-marketplace-wj2km\" (UID: \"2acd71c9-2eb5-49c2-9e2f-310b05980b3d\") " pod="openshift-marketplace/redhat-marketplace-wj2km" Nov 23 07:22:39 crc kubenswrapper[4559]: I1123 07:22:39.753042 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2acd71c9-2eb5-49c2-9e2f-310b05980b3d-catalog-content\") pod \"redhat-marketplace-wj2km\" (UID: \"2acd71c9-2eb5-49c2-9e2f-310b05980b3d\") " pod="openshift-marketplace/redhat-marketplace-wj2km" Nov 23 07:22:39 crc kubenswrapper[4559]: I1123 07:22:39.768531 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbp9w\" (UniqueName: \"kubernetes.io/projected/2acd71c9-2eb5-49c2-9e2f-310b05980b3d-kube-api-access-qbp9w\") pod \"redhat-marketplace-wj2km\" (UID: \"2acd71c9-2eb5-49c2-9e2f-310b05980b3d\") " pod="openshift-marketplace/redhat-marketplace-wj2km" Nov 23 07:22:39 crc kubenswrapper[4559]: I1123 07:22:39.971665 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wj2km" Nov 23 07:22:40 crc kubenswrapper[4559]: I1123 07:22:40.386332 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wj2km"] Nov 23 07:22:41 crc kubenswrapper[4559]: I1123 07:22:41.166988 4559 generic.go:334] "Generic (PLEG): container finished" podID="2acd71c9-2eb5-49c2-9e2f-310b05980b3d" containerID="3941c886feb628a9b942a58ebd3279b17de5b2d79cd2e32764d5cc0c4aa75e3c" exitCode=0 Nov 23 07:22:41 crc kubenswrapper[4559]: I1123 07:22:41.167094 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wj2km" event={"ID":"2acd71c9-2eb5-49c2-9e2f-310b05980b3d","Type":"ContainerDied","Data":"3941c886feb628a9b942a58ebd3279b17de5b2d79cd2e32764d5cc0c4aa75e3c"} Nov 23 07:22:41 crc kubenswrapper[4559]: I1123 07:22:41.167354 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wj2km" event={"ID":"2acd71c9-2eb5-49c2-9e2f-310b05980b3d","Type":"ContainerStarted","Data":"c70fdd629b813cfd32f52120352408fc1b281785825a7d2e8f5107e60972106c"} Nov 23 07:22:42 crc kubenswrapper[4559]: I1123 07:22:42.175803 4559 generic.go:334] "Generic (PLEG): container finished" podID="2acd71c9-2eb5-49c2-9e2f-310b05980b3d" containerID="e3442a31a9199137c971d7cb2a586d14e9a23496907dd236114fc319e4d1172e" exitCode=0 Nov 23 07:22:42 crc kubenswrapper[4559]: I1123 07:22:42.175910 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wj2km" event={"ID":"2acd71c9-2eb5-49c2-9e2f-310b05980b3d","Type":"ContainerDied","Data":"e3442a31a9199137c971d7cb2a586d14e9a23496907dd236114fc319e4d1172e"} Nov 23 07:22:43 crc kubenswrapper[4559]: I1123 07:22:43.184300 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wj2km" event={"ID":"2acd71c9-2eb5-49c2-9e2f-310b05980b3d","Type":"ContainerStarted","Data":"9bc9b49f433ba0e4c2ff9b673b952f06304c2f9059c668a16d29bfc081d8220b"} Nov 23 07:22:43 crc kubenswrapper[4559]: I1123 07:22:43.196717 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wj2km" podStartSLOduration=2.684195785 podStartE2EDuration="4.196701258s" podCreationTimestamp="2025-11-23 07:22:39 +0000 UTC" firstStartedPulling="2025-11-23 07:22:41.168807402 +0000 UTC m=+2263.190793016" lastFinishedPulling="2025-11-23 07:22:42.681312874 +0000 UTC m=+2264.703298489" observedRunningTime="2025-11-23 07:22:43.195923556 +0000 UTC m=+2265.217909169" watchObservedRunningTime="2025-11-23 07:22:43.196701258 +0000 UTC m=+2265.218686871" Nov 23 07:22:44 crc kubenswrapper[4559]: I1123 07:22:44.273729 4559 scope.go:117] "RemoveContainer" containerID="e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f" Nov 23 07:22:44 crc kubenswrapper[4559]: E1123 07:22:44.274251 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:22:49 crc kubenswrapper[4559]: I1123 07:22:49.972164 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wj2km" Nov 23 07:22:49 crc kubenswrapper[4559]: I1123 07:22:49.972592 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wj2km" Nov 23 07:22:50 crc kubenswrapper[4559]: I1123 07:22:50.002706 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wj2km" Nov 23 07:22:50 crc kubenswrapper[4559]: I1123 07:22:50.262857 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wj2km" Nov 23 07:22:50 crc kubenswrapper[4559]: I1123 07:22:50.297161 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wj2km"] Nov 23 07:22:52 crc kubenswrapper[4559]: I1123 07:22:52.242517 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wj2km" podUID="2acd71c9-2eb5-49c2-9e2f-310b05980b3d" containerName="registry-server" containerID="cri-o://9bc9b49f433ba0e4c2ff9b673b952f06304c2f9059c668a16d29bfc081d8220b" gracePeriod=2 Nov 23 07:22:52 crc kubenswrapper[4559]: I1123 07:22:52.615678 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wj2km" Nov 23 07:22:52 crc kubenswrapper[4559]: I1123 07:22:52.772377 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbp9w\" (UniqueName: \"kubernetes.io/projected/2acd71c9-2eb5-49c2-9e2f-310b05980b3d-kube-api-access-qbp9w\") pod \"2acd71c9-2eb5-49c2-9e2f-310b05980b3d\" (UID: \"2acd71c9-2eb5-49c2-9e2f-310b05980b3d\") " Nov 23 07:22:52 crc kubenswrapper[4559]: I1123 07:22:52.772426 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2acd71c9-2eb5-49c2-9e2f-310b05980b3d-utilities\") pod \"2acd71c9-2eb5-49c2-9e2f-310b05980b3d\" (UID: \"2acd71c9-2eb5-49c2-9e2f-310b05980b3d\") " Nov 23 07:22:52 crc kubenswrapper[4559]: I1123 07:22:52.772500 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2acd71c9-2eb5-49c2-9e2f-310b05980b3d-catalog-content\") pod \"2acd71c9-2eb5-49c2-9e2f-310b05980b3d\" (UID: \"2acd71c9-2eb5-49c2-9e2f-310b05980b3d\") " Nov 23 07:22:52 crc kubenswrapper[4559]: I1123 07:22:52.773118 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2acd71c9-2eb5-49c2-9e2f-310b05980b3d-utilities" (OuterVolumeSpecName: "utilities") pod "2acd71c9-2eb5-49c2-9e2f-310b05980b3d" (UID: "2acd71c9-2eb5-49c2-9e2f-310b05980b3d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:22:52 crc kubenswrapper[4559]: I1123 07:22:52.775083 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2acd71c9-2eb5-49c2-9e2f-310b05980b3d-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:22:52 crc kubenswrapper[4559]: I1123 07:22:52.777161 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2acd71c9-2eb5-49c2-9e2f-310b05980b3d-kube-api-access-qbp9w" (OuterVolumeSpecName: "kube-api-access-qbp9w") pod "2acd71c9-2eb5-49c2-9e2f-310b05980b3d" (UID: "2acd71c9-2eb5-49c2-9e2f-310b05980b3d"). InnerVolumeSpecName "kube-api-access-qbp9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:22:52 crc kubenswrapper[4559]: I1123 07:22:52.785245 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2acd71c9-2eb5-49c2-9e2f-310b05980b3d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2acd71c9-2eb5-49c2-9e2f-310b05980b3d" (UID: "2acd71c9-2eb5-49c2-9e2f-310b05980b3d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:22:52 crc kubenswrapper[4559]: I1123 07:22:52.876825 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbp9w\" (UniqueName: \"kubernetes.io/projected/2acd71c9-2eb5-49c2-9e2f-310b05980b3d-kube-api-access-qbp9w\") on node \"crc\" DevicePath \"\"" Nov 23 07:22:52 crc kubenswrapper[4559]: I1123 07:22:52.876849 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2acd71c9-2eb5-49c2-9e2f-310b05980b3d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:22:53 crc kubenswrapper[4559]: I1123 07:22:53.250975 4559 generic.go:334] "Generic (PLEG): container finished" podID="2acd71c9-2eb5-49c2-9e2f-310b05980b3d" containerID="9bc9b49f433ba0e4c2ff9b673b952f06304c2f9059c668a16d29bfc081d8220b" exitCode=0 Nov 23 07:22:53 crc kubenswrapper[4559]: I1123 07:22:53.251011 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wj2km" event={"ID":"2acd71c9-2eb5-49c2-9e2f-310b05980b3d","Type":"ContainerDied","Data":"9bc9b49f433ba0e4c2ff9b673b952f06304c2f9059c668a16d29bfc081d8220b"} Nov 23 07:22:53 crc kubenswrapper[4559]: I1123 07:22:53.251032 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wj2km" Nov 23 07:22:53 crc kubenswrapper[4559]: I1123 07:22:53.251049 4559 scope.go:117] "RemoveContainer" containerID="9bc9b49f433ba0e4c2ff9b673b952f06304c2f9059c668a16d29bfc081d8220b" Nov 23 07:22:53 crc kubenswrapper[4559]: I1123 07:22:53.251038 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wj2km" event={"ID":"2acd71c9-2eb5-49c2-9e2f-310b05980b3d","Type":"ContainerDied","Data":"c70fdd629b813cfd32f52120352408fc1b281785825a7d2e8f5107e60972106c"} Nov 23 07:22:53 crc kubenswrapper[4559]: I1123 07:22:53.268123 4559 scope.go:117] "RemoveContainer" containerID="e3442a31a9199137c971d7cb2a586d14e9a23496907dd236114fc319e4d1172e" Nov 23 07:22:53 crc kubenswrapper[4559]: I1123 07:22:53.274600 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wj2km"] Nov 23 07:22:53 crc kubenswrapper[4559]: I1123 07:22:53.280406 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wj2km"] Nov 23 07:22:53 crc kubenswrapper[4559]: I1123 07:22:53.297861 4559 scope.go:117] "RemoveContainer" containerID="3941c886feb628a9b942a58ebd3279b17de5b2d79cd2e32764d5cc0c4aa75e3c" Nov 23 07:22:53 crc kubenswrapper[4559]: I1123 07:22:53.320384 4559 scope.go:117] "RemoveContainer" containerID="9bc9b49f433ba0e4c2ff9b673b952f06304c2f9059c668a16d29bfc081d8220b" Nov 23 07:22:53 crc kubenswrapper[4559]: E1123 07:22:53.320761 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bc9b49f433ba0e4c2ff9b673b952f06304c2f9059c668a16d29bfc081d8220b\": container with ID starting with 9bc9b49f433ba0e4c2ff9b673b952f06304c2f9059c668a16d29bfc081d8220b not found: ID does not exist" containerID="9bc9b49f433ba0e4c2ff9b673b952f06304c2f9059c668a16d29bfc081d8220b" Nov 23 07:22:53 crc kubenswrapper[4559]: I1123 07:22:53.320789 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bc9b49f433ba0e4c2ff9b673b952f06304c2f9059c668a16d29bfc081d8220b"} err="failed to get container status \"9bc9b49f433ba0e4c2ff9b673b952f06304c2f9059c668a16d29bfc081d8220b\": rpc error: code = NotFound desc = could not find container \"9bc9b49f433ba0e4c2ff9b673b952f06304c2f9059c668a16d29bfc081d8220b\": container with ID starting with 9bc9b49f433ba0e4c2ff9b673b952f06304c2f9059c668a16d29bfc081d8220b not found: ID does not exist" Nov 23 07:22:53 crc kubenswrapper[4559]: I1123 07:22:53.320806 4559 scope.go:117] "RemoveContainer" containerID="e3442a31a9199137c971d7cb2a586d14e9a23496907dd236114fc319e4d1172e" Nov 23 07:22:53 crc kubenswrapper[4559]: E1123 07:22:53.320988 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3442a31a9199137c971d7cb2a586d14e9a23496907dd236114fc319e4d1172e\": container with ID starting with e3442a31a9199137c971d7cb2a586d14e9a23496907dd236114fc319e4d1172e not found: ID does not exist" containerID="e3442a31a9199137c971d7cb2a586d14e9a23496907dd236114fc319e4d1172e" Nov 23 07:22:53 crc kubenswrapper[4559]: I1123 07:22:53.321006 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3442a31a9199137c971d7cb2a586d14e9a23496907dd236114fc319e4d1172e"} err="failed to get container status \"e3442a31a9199137c971d7cb2a586d14e9a23496907dd236114fc319e4d1172e\": rpc error: code = NotFound desc = could not find container \"e3442a31a9199137c971d7cb2a586d14e9a23496907dd236114fc319e4d1172e\": container with ID starting with e3442a31a9199137c971d7cb2a586d14e9a23496907dd236114fc319e4d1172e not found: ID does not exist" Nov 23 07:22:53 crc kubenswrapper[4559]: I1123 07:22:53.321018 4559 scope.go:117] "RemoveContainer" containerID="3941c886feb628a9b942a58ebd3279b17de5b2d79cd2e32764d5cc0c4aa75e3c" Nov 23 07:22:53 crc kubenswrapper[4559]: E1123 07:22:53.321455 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3941c886feb628a9b942a58ebd3279b17de5b2d79cd2e32764d5cc0c4aa75e3c\": container with ID starting with 3941c886feb628a9b942a58ebd3279b17de5b2d79cd2e32764d5cc0c4aa75e3c not found: ID does not exist" containerID="3941c886feb628a9b942a58ebd3279b17de5b2d79cd2e32764d5cc0c4aa75e3c" Nov 23 07:22:53 crc kubenswrapper[4559]: I1123 07:22:53.321474 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3941c886feb628a9b942a58ebd3279b17de5b2d79cd2e32764d5cc0c4aa75e3c"} err="failed to get container status \"3941c886feb628a9b942a58ebd3279b17de5b2d79cd2e32764d5cc0c4aa75e3c\": rpc error: code = NotFound desc = could not find container \"3941c886feb628a9b942a58ebd3279b17de5b2d79cd2e32764d5cc0c4aa75e3c\": container with ID starting with 3941c886feb628a9b942a58ebd3279b17de5b2d79cd2e32764d5cc0c4aa75e3c not found: ID does not exist" Nov 23 07:22:54 crc kubenswrapper[4559]: I1123 07:22:54.281242 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2acd71c9-2eb5-49c2-9e2f-310b05980b3d" path="/var/lib/kubelet/pods/2acd71c9-2eb5-49c2-9e2f-310b05980b3d/volumes" Nov 23 07:22:55 crc kubenswrapper[4559]: I1123 07:22:55.273493 4559 scope.go:117] "RemoveContainer" containerID="e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f" Nov 23 07:22:55 crc kubenswrapper[4559]: E1123 07:22:55.273939 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:23:06 crc kubenswrapper[4559]: I1123 07:23:06.273239 4559 scope.go:117] "RemoveContainer" containerID="e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f" Nov 23 07:23:06 crc kubenswrapper[4559]: E1123 07:23:06.273829 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:23:17 crc kubenswrapper[4559]: I1123 07:23:17.273741 4559 scope.go:117] "RemoveContainer" containerID="e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f" Nov 23 07:23:17 crc kubenswrapper[4559]: E1123 07:23:17.274324 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:23:28 crc kubenswrapper[4559]: I1123 07:23:28.278501 4559 scope.go:117] "RemoveContainer" containerID="e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f" Nov 23 07:23:28 crc kubenswrapper[4559]: E1123 07:23:28.279094 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:23:43 crc kubenswrapper[4559]: I1123 07:23:43.273498 4559 scope.go:117] "RemoveContainer" containerID="e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f" Nov 23 07:23:43 crc kubenswrapper[4559]: E1123 07:23:43.274097 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:23:58 crc kubenswrapper[4559]: I1123 07:23:58.278323 4559 scope.go:117] "RemoveContainer" containerID="e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f" Nov 23 07:23:58 crc kubenswrapper[4559]: E1123 07:23:58.278941 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:24:09 crc kubenswrapper[4559]: I1123 07:24:09.273952 4559 scope.go:117] "RemoveContainer" containerID="e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f" Nov 23 07:24:09 crc kubenswrapper[4559]: E1123 07:24:09.274469 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:24:22 crc kubenswrapper[4559]: I1123 07:24:22.273982 4559 scope.go:117] "RemoveContainer" containerID="e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f" Nov 23 07:24:22 crc kubenswrapper[4559]: E1123 07:24:22.274572 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:24:33 crc kubenswrapper[4559]: I1123 07:24:33.273148 4559 scope.go:117] "RemoveContainer" containerID="e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f" Nov 23 07:24:33 crc kubenswrapper[4559]: E1123 07:24:33.273725 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:24:47 crc kubenswrapper[4559]: I1123 07:24:47.273116 4559 scope.go:117] "RemoveContainer" containerID="e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f" Nov 23 07:24:47 crc kubenswrapper[4559]: E1123 07:24:47.273694 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:24:59 crc kubenswrapper[4559]: I1123 07:24:59.274033 4559 scope.go:117] "RemoveContainer" containerID="e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f" Nov 23 07:24:59 crc kubenswrapper[4559]: E1123 07:24:59.274560 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:25:11 crc kubenswrapper[4559]: I1123 07:25:11.273676 4559 scope.go:117] "RemoveContainer" containerID="e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f" Nov 23 07:25:11 crc kubenswrapper[4559]: E1123 07:25:11.274397 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:25:24 crc kubenswrapper[4559]: I1123 07:25:24.274203 4559 scope.go:117] "RemoveContainer" containerID="e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f" Nov 23 07:25:24 crc kubenswrapper[4559]: E1123 07:25:24.275471 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:25:36 crc kubenswrapper[4559]: I1123 07:25:36.273226 4559 scope.go:117] "RemoveContainer" containerID="e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f" Nov 23 07:25:36 crc kubenswrapper[4559]: E1123 07:25:36.273940 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:25:49 crc kubenswrapper[4559]: I1123 07:25:49.968605 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l227w"] Nov 23 07:25:49 crc kubenswrapper[4559]: E1123 07:25:49.969404 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2acd71c9-2eb5-49c2-9e2f-310b05980b3d" containerName="registry-server" Nov 23 07:25:49 crc kubenswrapper[4559]: I1123 07:25:49.969418 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="2acd71c9-2eb5-49c2-9e2f-310b05980b3d" containerName="registry-server" Nov 23 07:25:49 crc kubenswrapper[4559]: E1123 07:25:49.969439 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2acd71c9-2eb5-49c2-9e2f-310b05980b3d" containerName="extract-content" Nov 23 07:25:49 crc kubenswrapper[4559]: I1123 07:25:49.969445 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="2acd71c9-2eb5-49c2-9e2f-310b05980b3d" containerName="extract-content" Nov 23 07:25:49 crc kubenswrapper[4559]: E1123 07:25:49.969454 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2acd71c9-2eb5-49c2-9e2f-310b05980b3d" containerName="extract-utilities" Nov 23 07:25:49 crc kubenswrapper[4559]: I1123 07:25:49.969460 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="2acd71c9-2eb5-49c2-9e2f-310b05980b3d" containerName="extract-utilities" Nov 23 07:25:49 crc kubenswrapper[4559]: I1123 07:25:49.969636 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="2acd71c9-2eb5-49c2-9e2f-310b05980b3d" containerName="registry-server" Nov 23 07:25:49 crc kubenswrapper[4559]: I1123 07:25:49.970769 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l227w" Nov 23 07:25:49 crc kubenswrapper[4559]: I1123 07:25:49.977160 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l227w"] Nov 23 07:25:49 crc kubenswrapper[4559]: I1123 07:25:49.994107 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7aa83d9b-23d7-4631-9d6d-8ab90441fbb2-catalog-content\") pod \"redhat-operators-l227w\" (UID: \"7aa83d9b-23d7-4631-9d6d-8ab90441fbb2\") " pod="openshift-marketplace/redhat-operators-l227w" Nov 23 07:25:49 crc kubenswrapper[4559]: I1123 07:25:49.994141 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8m4fw\" (UniqueName: \"kubernetes.io/projected/7aa83d9b-23d7-4631-9d6d-8ab90441fbb2-kube-api-access-8m4fw\") pod \"redhat-operators-l227w\" (UID: \"7aa83d9b-23d7-4631-9d6d-8ab90441fbb2\") " pod="openshift-marketplace/redhat-operators-l227w" Nov 23 07:25:49 crc kubenswrapper[4559]: I1123 07:25:49.994187 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7aa83d9b-23d7-4631-9d6d-8ab90441fbb2-utilities\") pod \"redhat-operators-l227w\" (UID: \"7aa83d9b-23d7-4631-9d6d-8ab90441fbb2\") " pod="openshift-marketplace/redhat-operators-l227w" Nov 23 07:25:50 crc kubenswrapper[4559]: I1123 07:25:50.095892 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7aa83d9b-23d7-4631-9d6d-8ab90441fbb2-utilities\") pod \"redhat-operators-l227w\" (UID: \"7aa83d9b-23d7-4631-9d6d-8ab90441fbb2\") " pod="openshift-marketplace/redhat-operators-l227w" Nov 23 07:25:50 crc kubenswrapper[4559]: I1123 07:25:50.096210 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7aa83d9b-23d7-4631-9d6d-8ab90441fbb2-catalog-content\") pod \"redhat-operators-l227w\" (UID: \"7aa83d9b-23d7-4631-9d6d-8ab90441fbb2\") " pod="openshift-marketplace/redhat-operators-l227w" Nov 23 07:25:50 crc kubenswrapper[4559]: I1123 07:25:50.096239 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8m4fw\" (UniqueName: \"kubernetes.io/projected/7aa83d9b-23d7-4631-9d6d-8ab90441fbb2-kube-api-access-8m4fw\") pod \"redhat-operators-l227w\" (UID: \"7aa83d9b-23d7-4631-9d6d-8ab90441fbb2\") " pod="openshift-marketplace/redhat-operators-l227w" Nov 23 07:25:50 crc kubenswrapper[4559]: I1123 07:25:50.096319 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7aa83d9b-23d7-4631-9d6d-8ab90441fbb2-utilities\") pod \"redhat-operators-l227w\" (UID: \"7aa83d9b-23d7-4631-9d6d-8ab90441fbb2\") " pod="openshift-marketplace/redhat-operators-l227w" Nov 23 07:25:50 crc kubenswrapper[4559]: I1123 07:25:50.096600 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7aa83d9b-23d7-4631-9d6d-8ab90441fbb2-catalog-content\") pod \"redhat-operators-l227w\" (UID: \"7aa83d9b-23d7-4631-9d6d-8ab90441fbb2\") " pod="openshift-marketplace/redhat-operators-l227w" Nov 23 07:25:50 crc kubenswrapper[4559]: I1123 07:25:50.116841 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8m4fw\" (UniqueName: \"kubernetes.io/projected/7aa83d9b-23d7-4631-9d6d-8ab90441fbb2-kube-api-access-8m4fw\") pod \"redhat-operators-l227w\" (UID: \"7aa83d9b-23d7-4631-9d6d-8ab90441fbb2\") " pod="openshift-marketplace/redhat-operators-l227w" Nov 23 07:25:50 crc kubenswrapper[4559]: I1123 07:25:50.274000 4559 scope.go:117] "RemoveContainer" containerID="e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f" Nov 23 07:25:50 crc kubenswrapper[4559]: E1123 07:25:50.274347 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:25:50 crc kubenswrapper[4559]: I1123 07:25:50.300001 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l227w" Nov 23 07:25:50 crc kubenswrapper[4559]: I1123 07:25:50.705136 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l227w"] Nov 23 07:25:51 crc kubenswrapper[4559]: I1123 07:25:51.360491 4559 generic.go:334] "Generic (PLEG): container finished" podID="7aa83d9b-23d7-4631-9d6d-8ab90441fbb2" containerID="7b3a4564773f4e1fa4d97972f5d7da64ee33e7d0e7b678166278cef2f12bc9e2" exitCode=0 Nov 23 07:25:51 crc kubenswrapper[4559]: I1123 07:25:51.360538 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l227w" event={"ID":"7aa83d9b-23d7-4631-9d6d-8ab90441fbb2","Type":"ContainerDied","Data":"7b3a4564773f4e1fa4d97972f5d7da64ee33e7d0e7b678166278cef2f12bc9e2"} Nov 23 07:25:51 crc kubenswrapper[4559]: I1123 07:25:51.360831 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l227w" event={"ID":"7aa83d9b-23d7-4631-9d6d-8ab90441fbb2","Type":"ContainerStarted","Data":"47aee9bdd9e61bfe7808c9e4bd0a3bdf3c4595a7204f54ad3ad799f9c9295b16"} Nov 23 07:25:51 crc kubenswrapper[4559]: I1123 07:25:51.362100 4559 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 07:25:52 crc kubenswrapper[4559]: I1123 07:25:52.374548 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l227w" event={"ID":"7aa83d9b-23d7-4631-9d6d-8ab90441fbb2","Type":"ContainerStarted","Data":"45f326914b1bcd98315da3dfe42fd8a7e165484714de94d4482f805e783f849d"} Nov 23 07:25:54 crc kubenswrapper[4559]: I1123 07:25:54.388577 4559 generic.go:334] "Generic (PLEG): container finished" podID="7aa83d9b-23d7-4631-9d6d-8ab90441fbb2" containerID="45f326914b1bcd98315da3dfe42fd8a7e165484714de94d4482f805e783f849d" exitCode=0 Nov 23 07:25:54 crc kubenswrapper[4559]: I1123 07:25:54.388665 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l227w" event={"ID":"7aa83d9b-23d7-4631-9d6d-8ab90441fbb2","Type":"ContainerDied","Data":"45f326914b1bcd98315da3dfe42fd8a7e165484714de94d4482f805e783f849d"} Nov 23 07:25:55 crc kubenswrapper[4559]: I1123 07:25:55.398319 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l227w" event={"ID":"7aa83d9b-23d7-4631-9d6d-8ab90441fbb2","Type":"ContainerStarted","Data":"3dd508d1d0ca8646211548e2c25707e54bf937344c259947d4d30435bad14d4b"} Nov 23 07:26:00 crc kubenswrapper[4559]: I1123 07:26:00.300806 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l227w" Nov 23 07:26:00 crc kubenswrapper[4559]: I1123 07:26:00.301314 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l227w" Nov 23 07:26:00 crc kubenswrapper[4559]: I1123 07:26:00.331927 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l227w" Nov 23 07:26:00 crc kubenswrapper[4559]: I1123 07:26:00.345590 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l227w" podStartSLOduration=7.774988037 podStartE2EDuration="11.345575138s" podCreationTimestamp="2025-11-23 07:25:49 +0000 UTC" firstStartedPulling="2025-11-23 07:25:51.361907044 +0000 UTC m=+2453.383892658" lastFinishedPulling="2025-11-23 07:25:54.932494144 +0000 UTC m=+2456.954479759" observedRunningTime="2025-11-23 07:25:55.415224466 +0000 UTC m=+2457.437210080" watchObservedRunningTime="2025-11-23 07:26:00.345575138 +0000 UTC m=+2462.367560752" Nov 23 07:26:00 crc kubenswrapper[4559]: I1123 07:26:00.458932 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l227w" Nov 23 07:26:00 crc kubenswrapper[4559]: I1123 07:26:00.557122 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l227w"] Nov 23 07:26:02 crc kubenswrapper[4559]: I1123 07:26:02.442218 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-l227w" podUID="7aa83d9b-23d7-4631-9d6d-8ab90441fbb2" containerName="registry-server" containerID="cri-o://3dd508d1d0ca8646211548e2c25707e54bf937344c259947d4d30435bad14d4b" gracePeriod=2 Nov 23 07:26:02 crc kubenswrapper[4559]: I1123 07:26:02.776918 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l227w" Nov 23 07:26:02 crc kubenswrapper[4559]: I1123 07:26:02.897413 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7aa83d9b-23d7-4631-9d6d-8ab90441fbb2-utilities\") pod \"7aa83d9b-23d7-4631-9d6d-8ab90441fbb2\" (UID: \"7aa83d9b-23d7-4631-9d6d-8ab90441fbb2\") " Nov 23 07:26:02 crc kubenswrapper[4559]: I1123 07:26:02.897721 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8m4fw\" (UniqueName: \"kubernetes.io/projected/7aa83d9b-23d7-4631-9d6d-8ab90441fbb2-kube-api-access-8m4fw\") pod \"7aa83d9b-23d7-4631-9d6d-8ab90441fbb2\" (UID: \"7aa83d9b-23d7-4631-9d6d-8ab90441fbb2\") " Nov 23 07:26:02 crc kubenswrapper[4559]: I1123 07:26:02.897768 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7aa83d9b-23d7-4631-9d6d-8ab90441fbb2-catalog-content\") pod \"7aa83d9b-23d7-4631-9d6d-8ab90441fbb2\" (UID: \"7aa83d9b-23d7-4631-9d6d-8ab90441fbb2\") " Nov 23 07:26:02 crc kubenswrapper[4559]: I1123 07:26:02.898063 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7aa83d9b-23d7-4631-9d6d-8ab90441fbb2-utilities" (OuterVolumeSpecName: "utilities") pod "7aa83d9b-23d7-4631-9d6d-8ab90441fbb2" (UID: "7aa83d9b-23d7-4631-9d6d-8ab90441fbb2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:26:02 crc kubenswrapper[4559]: I1123 07:26:02.898444 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7aa83d9b-23d7-4631-9d6d-8ab90441fbb2-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:26:02 crc kubenswrapper[4559]: I1123 07:26:02.903744 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7aa83d9b-23d7-4631-9d6d-8ab90441fbb2-kube-api-access-8m4fw" (OuterVolumeSpecName: "kube-api-access-8m4fw") pod "7aa83d9b-23d7-4631-9d6d-8ab90441fbb2" (UID: "7aa83d9b-23d7-4631-9d6d-8ab90441fbb2"). InnerVolumeSpecName "kube-api-access-8m4fw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:26:02 crc kubenswrapper[4559]: I1123 07:26:02.966995 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7aa83d9b-23d7-4631-9d6d-8ab90441fbb2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7aa83d9b-23d7-4631-9d6d-8ab90441fbb2" (UID: "7aa83d9b-23d7-4631-9d6d-8ab90441fbb2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:26:02 crc kubenswrapper[4559]: I1123 07:26:02.999653 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8m4fw\" (UniqueName: \"kubernetes.io/projected/7aa83d9b-23d7-4631-9d6d-8ab90441fbb2-kube-api-access-8m4fw\") on node \"crc\" DevicePath \"\"" Nov 23 07:26:02 crc kubenswrapper[4559]: I1123 07:26:02.999682 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7aa83d9b-23d7-4631-9d6d-8ab90441fbb2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:26:03 crc kubenswrapper[4559]: I1123 07:26:03.273685 4559 scope.go:117] "RemoveContainer" containerID="e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f" Nov 23 07:26:03 crc kubenswrapper[4559]: I1123 07:26:03.451967 4559 generic.go:334] "Generic (PLEG): container finished" podID="7aa83d9b-23d7-4631-9d6d-8ab90441fbb2" containerID="3dd508d1d0ca8646211548e2c25707e54bf937344c259947d4d30435bad14d4b" exitCode=0 Nov 23 07:26:03 crc kubenswrapper[4559]: I1123 07:26:03.452148 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l227w" event={"ID":"7aa83d9b-23d7-4631-9d6d-8ab90441fbb2","Type":"ContainerDied","Data":"3dd508d1d0ca8646211548e2c25707e54bf937344c259947d4d30435bad14d4b"} Nov 23 07:26:03 crc kubenswrapper[4559]: I1123 07:26:03.452213 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l227w" event={"ID":"7aa83d9b-23d7-4631-9d6d-8ab90441fbb2","Type":"ContainerDied","Data":"47aee9bdd9e61bfe7808c9e4bd0a3bdf3c4595a7204f54ad3ad799f9c9295b16"} Nov 23 07:26:03 crc kubenswrapper[4559]: I1123 07:26:03.452233 4559 scope.go:117] "RemoveContainer" containerID="3dd508d1d0ca8646211548e2c25707e54bf937344c259947d4d30435bad14d4b" Nov 23 07:26:03 crc kubenswrapper[4559]: I1123 07:26:03.452269 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l227w" Nov 23 07:26:03 crc kubenswrapper[4559]: I1123 07:26:03.455578 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerStarted","Data":"3b5b7bce47b314a27a0c4809c674e865bcfcc7e7b6194e23f9fb9dd9bd126d02"} Nov 23 07:26:03 crc kubenswrapper[4559]: I1123 07:26:03.470767 4559 scope.go:117] "RemoveContainer" containerID="45f326914b1bcd98315da3dfe42fd8a7e165484714de94d4482f805e783f849d" Nov 23 07:26:03 crc kubenswrapper[4559]: I1123 07:26:03.510035 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l227w"] Nov 23 07:26:03 crc kubenswrapper[4559]: I1123 07:26:03.510083 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-l227w"] Nov 23 07:26:03 crc kubenswrapper[4559]: I1123 07:26:03.518163 4559 scope.go:117] "RemoveContainer" containerID="7b3a4564773f4e1fa4d97972f5d7da64ee33e7d0e7b678166278cef2f12bc9e2" Nov 23 07:26:03 crc kubenswrapper[4559]: I1123 07:26:03.537266 4559 scope.go:117] "RemoveContainer" containerID="3dd508d1d0ca8646211548e2c25707e54bf937344c259947d4d30435bad14d4b" Nov 23 07:26:03 crc kubenswrapper[4559]: E1123 07:26:03.537754 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3dd508d1d0ca8646211548e2c25707e54bf937344c259947d4d30435bad14d4b\": container with ID starting with 3dd508d1d0ca8646211548e2c25707e54bf937344c259947d4d30435bad14d4b not found: ID does not exist" containerID="3dd508d1d0ca8646211548e2c25707e54bf937344c259947d4d30435bad14d4b" Nov 23 07:26:03 crc kubenswrapper[4559]: I1123 07:26:03.537800 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dd508d1d0ca8646211548e2c25707e54bf937344c259947d4d30435bad14d4b"} err="failed to get container status \"3dd508d1d0ca8646211548e2c25707e54bf937344c259947d4d30435bad14d4b\": rpc error: code = NotFound desc = could not find container \"3dd508d1d0ca8646211548e2c25707e54bf937344c259947d4d30435bad14d4b\": container with ID starting with 3dd508d1d0ca8646211548e2c25707e54bf937344c259947d4d30435bad14d4b not found: ID does not exist" Nov 23 07:26:03 crc kubenswrapper[4559]: I1123 07:26:03.537826 4559 scope.go:117] "RemoveContainer" containerID="45f326914b1bcd98315da3dfe42fd8a7e165484714de94d4482f805e783f849d" Nov 23 07:26:03 crc kubenswrapper[4559]: E1123 07:26:03.538188 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45f326914b1bcd98315da3dfe42fd8a7e165484714de94d4482f805e783f849d\": container with ID starting with 45f326914b1bcd98315da3dfe42fd8a7e165484714de94d4482f805e783f849d not found: ID does not exist" containerID="45f326914b1bcd98315da3dfe42fd8a7e165484714de94d4482f805e783f849d" Nov 23 07:26:03 crc kubenswrapper[4559]: I1123 07:26:03.538227 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45f326914b1bcd98315da3dfe42fd8a7e165484714de94d4482f805e783f849d"} err="failed to get container status \"45f326914b1bcd98315da3dfe42fd8a7e165484714de94d4482f805e783f849d\": rpc error: code = NotFound desc = could not find container \"45f326914b1bcd98315da3dfe42fd8a7e165484714de94d4482f805e783f849d\": container with ID starting with 45f326914b1bcd98315da3dfe42fd8a7e165484714de94d4482f805e783f849d not found: ID does not exist" Nov 23 07:26:03 crc kubenswrapper[4559]: I1123 07:26:03.538254 4559 scope.go:117] "RemoveContainer" containerID="7b3a4564773f4e1fa4d97972f5d7da64ee33e7d0e7b678166278cef2f12bc9e2" Nov 23 07:26:03 crc kubenswrapper[4559]: E1123 07:26:03.538542 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b3a4564773f4e1fa4d97972f5d7da64ee33e7d0e7b678166278cef2f12bc9e2\": container with ID starting with 7b3a4564773f4e1fa4d97972f5d7da64ee33e7d0e7b678166278cef2f12bc9e2 not found: ID does not exist" containerID="7b3a4564773f4e1fa4d97972f5d7da64ee33e7d0e7b678166278cef2f12bc9e2" Nov 23 07:26:03 crc kubenswrapper[4559]: I1123 07:26:03.538566 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b3a4564773f4e1fa4d97972f5d7da64ee33e7d0e7b678166278cef2f12bc9e2"} err="failed to get container status \"7b3a4564773f4e1fa4d97972f5d7da64ee33e7d0e7b678166278cef2f12bc9e2\": rpc error: code = NotFound desc = could not find container \"7b3a4564773f4e1fa4d97972f5d7da64ee33e7d0e7b678166278cef2f12bc9e2\": container with ID starting with 7b3a4564773f4e1fa4d97972f5d7da64ee33e7d0e7b678166278cef2f12bc9e2 not found: ID does not exist" Nov 23 07:26:04 crc kubenswrapper[4559]: I1123 07:26:04.281300 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7aa83d9b-23d7-4631-9d6d-8ab90441fbb2" path="/var/lib/kubelet/pods/7aa83d9b-23d7-4631-9d6d-8ab90441fbb2/volumes" Nov 23 07:28:26 crc kubenswrapper[4559]: I1123 07:28:26.167408 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:28:26 crc kubenswrapper[4559]: I1123 07:28:26.167869 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:28:56 crc kubenswrapper[4559]: I1123 07:28:56.167380 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:28:56 crc kubenswrapper[4559]: I1123 07:28:56.167854 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:29:26 crc kubenswrapper[4559]: I1123 07:29:26.167330 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:29:26 crc kubenswrapper[4559]: I1123 07:29:26.168037 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:29:26 crc kubenswrapper[4559]: I1123 07:29:26.168096 4559 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" Nov 23 07:29:26 crc kubenswrapper[4559]: I1123 07:29:26.168917 4559 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3b5b7bce47b314a27a0c4809c674e865bcfcc7e7b6194e23f9fb9dd9bd126d02"} pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 07:29:26 crc kubenswrapper[4559]: I1123 07:29:26.168969 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" containerID="cri-o://3b5b7bce47b314a27a0c4809c674e865bcfcc7e7b6194e23f9fb9dd9bd126d02" gracePeriod=600 Nov 23 07:29:26 crc kubenswrapper[4559]: I1123 07:29:26.760038 4559 generic.go:334] "Generic (PLEG): container finished" podID="4731beee-0cac-4189-8a70-743b0b709095" containerID="3b5b7bce47b314a27a0c4809c674e865bcfcc7e7b6194e23f9fb9dd9bd126d02" exitCode=0 Nov 23 07:29:26 crc kubenswrapper[4559]: I1123 07:29:26.760323 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerDied","Data":"3b5b7bce47b314a27a0c4809c674e865bcfcc7e7b6194e23f9fb9dd9bd126d02"} Nov 23 07:29:26 crc kubenswrapper[4559]: I1123 07:29:26.760349 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerStarted","Data":"c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4"} Nov 23 07:29:26 crc kubenswrapper[4559]: I1123 07:29:26.760365 4559 scope.go:117] "RemoveContainer" containerID="e6934c172298c610eb1a9026cc51308c1e8025159c624b13aae23c5753546a3f" Nov 23 07:29:27 crc kubenswrapper[4559]: I1123 07:29:27.770592 4559 generic.go:334] "Generic (PLEG): container finished" podID="fb299f6e-9a30-4d29-95db-a0cbceb573ea" containerID="3fb16f9f3381f4ad7ea54d90253d1175d802011895e8e41536ac004dae8676da" exitCode=0 Nov 23 07:29:27 crc kubenswrapper[4559]: I1123 07:29:27.770683 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"fb299f6e-9a30-4d29-95db-a0cbceb573ea","Type":"ContainerDied","Data":"3fb16f9f3381f4ad7ea54d90253d1175d802011895e8e41536ac004dae8676da"} Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.046719 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.119053 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb299f6e-9a30-4d29-95db-a0cbceb573ea-ssh-key\") pod \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.119095 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/fb299f6e-9a30-4d29-95db-a0cbceb573ea-test-operator-ephemeral-temporary\") pod \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.119192 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fb299f6e-9a30-4d29-95db-a0cbceb573ea-openstack-config\") pod \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.119210 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47ngx\" (UniqueName: \"kubernetes.io/projected/fb299f6e-9a30-4d29-95db-a0cbceb573ea-kube-api-access-47ngx\") pod \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.119241 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb299f6e-9a30-4d29-95db-a0cbceb573ea-config-data\") pod \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.119274 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.119295 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/fb299f6e-9a30-4d29-95db-a0cbceb573ea-test-operator-ephemeral-workdir\") pod \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.119346 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fb299f6e-9a30-4d29-95db-a0cbceb573ea-openstack-config-secret\") pod \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.119389 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/fb299f6e-9a30-4d29-95db-a0cbceb573ea-ca-certs\") pod \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\" (UID: \"fb299f6e-9a30-4d29-95db-a0cbceb573ea\") " Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.119705 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb299f6e-9a30-4d29-95db-a0cbceb573ea-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "fb299f6e-9a30-4d29-95db-a0cbceb573ea" (UID: "fb299f6e-9a30-4d29-95db-a0cbceb573ea"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.120080 4559 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/fb299f6e-9a30-4d29-95db-a0cbceb573ea-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.120289 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb299f6e-9a30-4d29-95db-a0cbceb573ea-config-data" (OuterVolumeSpecName: "config-data") pod "fb299f6e-9a30-4d29-95db-a0cbceb573ea" (UID: "fb299f6e-9a30-4d29-95db-a0cbceb573ea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.122576 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb299f6e-9a30-4d29-95db-a0cbceb573ea-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "fb299f6e-9a30-4d29-95db-a0cbceb573ea" (UID: "fb299f6e-9a30-4d29-95db-a0cbceb573ea"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.127542 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb299f6e-9a30-4d29-95db-a0cbceb573ea-kube-api-access-47ngx" (OuterVolumeSpecName: "kube-api-access-47ngx") pod "fb299f6e-9a30-4d29-95db-a0cbceb573ea" (UID: "fb299f6e-9a30-4d29-95db-a0cbceb573ea"). InnerVolumeSpecName "kube-api-access-47ngx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.129250 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "test-operator-logs") pod "fb299f6e-9a30-4d29-95db-a0cbceb573ea" (UID: "fb299f6e-9a30-4d29-95db-a0cbceb573ea"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.145284 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb299f6e-9a30-4d29-95db-a0cbceb573ea-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "fb299f6e-9a30-4d29-95db-a0cbceb573ea" (UID: "fb299f6e-9a30-4d29-95db-a0cbceb573ea"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.147028 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb299f6e-9a30-4d29-95db-a0cbceb573ea-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "fb299f6e-9a30-4d29-95db-a0cbceb573ea" (UID: "fb299f6e-9a30-4d29-95db-a0cbceb573ea"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.152389 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb299f6e-9a30-4d29-95db-a0cbceb573ea-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fb299f6e-9a30-4d29-95db-a0cbceb573ea" (UID: "fb299f6e-9a30-4d29-95db-a0cbceb573ea"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.160940 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb299f6e-9a30-4d29-95db-a0cbceb573ea-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "fb299f6e-9a30-4d29-95db-a0cbceb573ea" (UID: "fb299f6e-9a30-4d29-95db-a0cbceb573ea"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.220967 4559 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb299f6e-9a30-4d29-95db-a0cbceb573ea-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.220996 4559 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fb299f6e-9a30-4d29-95db-a0cbceb573ea-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.221008 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47ngx\" (UniqueName: \"kubernetes.io/projected/fb299f6e-9a30-4d29-95db-a0cbceb573ea-kube-api-access-47ngx\") on node \"crc\" DevicePath \"\"" Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.221018 4559 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb299f6e-9a30-4d29-95db-a0cbceb573ea-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.221044 4559 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.221057 4559 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/fb299f6e-9a30-4d29-95db-a0cbceb573ea-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.221066 4559 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fb299f6e-9a30-4d29-95db-a0cbceb573ea-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.221082 4559 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/fb299f6e-9a30-4d29-95db-a0cbceb573ea-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.234792 4559 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.321812 4559 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.783471 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"fb299f6e-9a30-4d29-95db-a0cbceb573ea","Type":"ContainerDied","Data":"620011a02e5d17c29b5afd433ca1db95a9b42700270ae3b6a76ea735f6879236"} Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.783510 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="620011a02e5d17c29b5afd433ca1db95a9b42700270ae3b6a76ea735f6879236" Nov 23 07:29:29 crc kubenswrapper[4559]: I1123 07:29:29.783558 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 23 07:29:37 crc kubenswrapper[4559]: I1123 07:29:37.636682 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 23 07:29:37 crc kubenswrapper[4559]: E1123 07:29:37.637421 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7aa83d9b-23d7-4631-9d6d-8ab90441fbb2" containerName="extract-utilities" Nov 23 07:29:37 crc kubenswrapper[4559]: I1123 07:29:37.637432 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="7aa83d9b-23d7-4631-9d6d-8ab90441fbb2" containerName="extract-utilities" Nov 23 07:29:37 crc kubenswrapper[4559]: E1123 07:29:37.637459 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb299f6e-9a30-4d29-95db-a0cbceb573ea" containerName="tempest-tests-tempest-tests-runner" Nov 23 07:29:37 crc kubenswrapper[4559]: I1123 07:29:37.637464 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb299f6e-9a30-4d29-95db-a0cbceb573ea" containerName="tempest-tests-tempest-tests-runner" Nov 23 07:29:37 crc kubenswrapper[4559]: E1123 07:29:37.637474 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7aa83d9b-23d7-4631-9d6d-8ab90441fbb2" containerName="extract-content" Nov 23 07:29:37 crc kubenswrapper[4559]: I1123 07:29:37.637480 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="7aa83d9b-23d7-4631-9d6d-8ab90441fbb2" containerName="extract-content" Nov 23 07:29:37 crc kubenswrapper[4559]: E1123 07:29:37.637494 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7aa83d9b-23d7-4631-9d6d-8ab90441fbb2" containerName="registry-server" Nov 23 07:29:37 crc kubenswrapper[4559]: I1123 07:29:37.637499 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="7aa83d9b-23d7-4631-9d6d-8ab90441fbb2" containerName="registry-server" Nov 23 07:29:37 crc kubenswrapper[4559]: I1123 07:29:37.637670 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="7aa83d9b-23d7-4631-9d6d-8ab90441fbb2" containerName="registry-server" Nov 23 07:29:37 crc kubenswrapper[4559]: I1123 07:29:37.637693 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb299f6e-9a30-4d29-95db-a0cbceb573ea" containerName="tempest-tests-tempest-tests-runner" Nov 23 07:29:37 crc kubenswrapper[4559]: I1123 07:29:37.638247 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 23 07:29:37 crc kubenswrapper[4559]: I1123 07:29:37.640614 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-pdp84" Nov 23 07:29:37 crc kubenswrapper[4559]: I1123 07:29:37.646792 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 23 07:29:37 crc kubenswrapper[4559]: I1123 07:29:37.760509 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9cf5\" (UniqueName: \"kubernetes.io/projected/83a546c2-d73a-40cd-aae5-22eb4aef5268-kube-api-access-t9cf5\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"83a546c2-d73a-40cd-aae5-22eb4aef5268\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 23 07:29:37 crc kubenswrapper[4559]: I1123 07:29:37.760657 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"83a546c2-d73a-40cd-aae5-22eb4aef5268\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 23 07:29:37 crc kubenswrapper[4559]: I1123 07:29:37.862783 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9cf5\" (UniqueName: \"kubernetes.io/projected/83a546c2-d73a-40cd-aae5-22eb4aef5268-kube-api-access-t9cf5\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"83a546c2-d73a-40cd-aae5-22eb4aef5268\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 23 07:29:37 crc kubenswrapper[4559]: I1123 07:29:37.862888 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"83a546c2-d73a-40cd-aae5-22eb4aef5268\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 23 07:29:37 crc kubenswrapper[4559]: I1123 07:29:37.863172 4559 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"83a546c2-d73a-40cd-aae5-22eb4aef5268\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 23 07:29:37 crc kubenswrapper[4559]: I1123 07:29:37.877815 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9cf5\" (UniqueName: \"kubernetes.io/projected/83a546c2-d73a-40cd-aae5-22eb4aef5268-kube-api-access-t9cf5\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"83a546c2-d73a-40cd-aae5-22eb4aef5268\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 23 07:29:37 crc kubenswrapper[4559]: I1123 07:29:37.881417 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"83a546c2-d73a-40cd-aae5-22eb4aef5268\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 23 07:29:37 crc kubenswrapper[4559]: I1123 07:29:37.954773 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 23 07:29:38 crc kubenswrapper[4559]: I1123 07:29:38.315705 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 23 07:29:38 crc kubenswrapper[4559]: I1123 07:29:38.844053 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"83a546c2-d73a-40cd-aae5-22eb4aef5268","Type":"ContainerStarted","Data":"2ccfdab40b4d3f25fc2bddc02a58e3f2022bfe5eda003b9301afaa06a1001f46"} Nov 23 07:29:39 crc kubenswrapper[4559]: I1123 07:29:39.851179 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"83a546c2-d73a-40cd-aae5-22eb4aef5268","Type":"ContainerStarted","Data":"3ddeaa25949ea120f872edb22970364f31ef3f9a4e8a578fcb6246f232530a35"} Nov 23 07:29:39 crc kubenswrapper[4559]: I1123 07:29:39.865406 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.100685281 podStartE2EDuration="2.865390524s" podCreationTimestamp="2025-11-23 07:29:37 +0000 UTC" firstStartedPulling="2025-11-23 07:29:38.322793418 +0000 UTC m=+2680.344779032" lastFinishedPulling="2025-11-23 07:29:39.087498651 +0000 UTC m=+2681.109484275" observedRunningTime="2025-11-23 07:29:39.859635185 +0000 UTC m=+2681.881620798" watchObservedRunningTime="2025-11-23 07:29:39.865390524 +0000 UTC m=+2681.887376138" Nov 23 07:29:58 crc kubenswrapper[4559]: I1123 07:29:58.188612 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-m8wzp/must-gather-8nhbp"] Nov 23 07:29:58 crc kubenswrapper[4559]: I1123 07:29:58.193014 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m8wzp/must-gather-8nhbp" Nov 23 07:29:58 crc kubenswrapper[4559]: I1123 07:29:58.196179 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-m8wzp"/"kube-root-ca.crt" Nov 23 07:29:58 crc kubenswrapper[4559]: I1123 07:29:58.196381 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-m8wzp"/"openshift-service-ca.crt" Nov 23 07:29:58 crc kubenswrapper[4559]: I1123 07:29:58.205249 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-m8wzp/must-gather-8nhbp"] Nov 23 07:29:58 crc kubenswrapper[4559]: I1123 07:29:58.205689 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-m8wzp"/"default-dockercfg-xqpd6" Nov 23 07:29:58 crc kubenswrapper[4559]: I1123 07:29:58.355890 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/13525b40-ea0a-4894-8622-e017b4f34397-must-gather-output\") pod \"must-gather-8nhbp\" (UID: \"13525b40-ea0a-4894-8622-e017b4f34397\") " pod="openshift-must-gather-m8wzp/must-gather-8nhbp" Nov 23 07:29:58 crc kubenswrapper[4559]: I1123 07:29:58.355952 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6b8z\" (UniqueName: \"kubernetes.io/projected/13525b40-ea0a-4894-8622-e017b4f34397-kube-api-access-q6b8z\") pod \"must-gather-8nhbp\" (UID: \"13525b40-ea0a-4894-8622-e017b4f34397\") " pod="openshift-must-gather-m8wzp/must-gather-8nhbp" Nov 23 07:29:58 crc kubenswrapper[4559]: I1123 07:29:58.457283 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/13525b40-ea0a-4894-8622-e017b4f34397-must-gather-output\") pod \"must-gather-8nhbp\" (UID: \"13525b40-ea0a-4894-8622-e017b4f34397\") " pod="openshift-must-gather-m8wzp/must-gather-8nhbp" Nov 23 07:29:58 crc kubenswrapper[4559]: I1123 07:29:58.457478 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6b8z\" (UniqueName: \"kubernetes.io/projected/13525b40-ea0a-4894-8622-e017b4f34397-kube-api-access-q6b8z\") pod \"must-gather-8nhbp\" (UID: \"13525b40-ea0a-4894-8622-e017b4f34397\") " pod="openshift-must-gather-m8wzp/must-gather-8nhbp" Nov 23 07:29:58 crc kubenswrapper[4559]: I1123 07:29:58.457637 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/13525b40-ea0a-4894-8622-e017b4f34397-must-gather-output\") pod \"must-gather-8nhbp\" (UID: \"13525b40-ea0a-4894-8622-e017b4f34397\") " pod="openshift-must-gather-m8wzp/must-gather-8nhbp" Nov 23 07:29:58 crc kubenswrapper[4559]: I1123 07:29:58.469239 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-m8wzp"/"kube-root-ca.crt" Nov 23 07:29:58 crc kubenswrapper[4559]: I1123 07:29:58.479070 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-m8wzp"/"openshift-service-ca.crt" Nov 23 07:29:58 crc kubenswrapper[4559]: I1123 07:29:58.493762 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6b8z\" (UniqueName: \"kubernetes.io/projected/13525b40-ea0a-4894-8622-e017b4f34397-kube-api-access-q6b8z\") pod \"must-gather-8nhbp\" (UID: \"13525b40-ea0a-4894-8622-e017b4f34397\") " pod="openshift-must-gather-m8wzp/must-gather-8nhbp" Nov 23 07:29:58 crc kubenswrapper[4559]: I1123 07:29:58.510066 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-m8wzp"/"default-dockercfg-xqpd6" Nov 23 07:29:58 crc kubenswrapper[4559]: I1123 07:29:58.519019 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m8wzp/must-gather-8nhbp" Nov 23 07:29:58 crc kubenswrapper[4559]: I1123 07:29:58.889620 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-m8wzp/must-gather-8nhbp"] Nov 23 07:29:58 crc kubenswrapper[4559]: W1123 07:29:58.898202 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13525b40_ea0a_4894_8622_e017b4f34397.slice/crio-f3d9c7eea2e817698e9e3add68538b28605898cefae4b0acffa215b966857e9c WatchSource:0}: Error finding container f3d9c7eea2e817698e9e3add68538b28605898cefae4b0acffa215b966857e9c: Status 404 returned error can't find the container with id f3d9c7eea2e817698e9e3add68538b28605898cefae4b0acffa215b966857e9c Nov 23 07:29:58 crc kubenswrapper[4559]: I1123 07:29:58.966330 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m8wzp/must-gather-8nhbp" event={"ID":"13525b40-ea0a-4894-8622-e017b4f34397","Type":"ContainerStarted","Data":"f3d9c7eea2e817698e9e3add68538b28605898cefae4b0acffa215b966857e9c"} Nov 23 07:30:00 crc kubenswrapper[4559]: I1123 07:30:00.137618 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398050-hghk6"] Nov 23 07:30:00 crc kubenswrapper[4559]: I1123 07:30:00.139220 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-hghk6" Nov 23 07:30:00 crc kubenswrapper[4559]: I1123 07:30:00.141142 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 23 07:30:00 crc kubenswrapper[4559]: I1123 07:30:00.141178 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 23 07:30:00 crc kubenswrapper[4559]: I1123 07:30:00.144949 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398050-hghk6"] Nov 23 07:30:00 crc kubenswrapper[4559]: I1123 07:30:00.188904 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5znm9\" (UniqueName: \"kubernetes.io/projected/edf3373e-6c91-419a-91b8-3cf9cd9e9725-kube-api-access-5znm9\") pod \"collect-profiles-29398050-hghk6\" (UID: \"edf3373e-6c91-419a-91b8-3cf9cd9e9725\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-hghk6" Nov 23 07:30:00 crc kubenswrapper[4559]: I1123 07:30:00.189117 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/edf3373e-6c91-419a-91b8-3cf9cd9e9725-config-volume\") pod \"collect-profiles-29398050-hghk6\" (UID: \"edf3373e-6c91-419a-91b8-3cf9cd9e9725\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-hghk6" Nov 23 07:30:00 crc kubenswrapper[4559]: I1123 07:30:00.189297 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/edf3373e-6c91-419a-91b8-3cf9cd9e9725-secret-volume\") pod \"collect-profiles-29398050-hghk6\" (UID: \"edf3373e-6c91-419a-91b8-3cf9cd9e9725\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-hghk6" Nov 23 07:30:00 crc kubenswrapper[4559]: I1123 07:30:00.290769 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5znm9\" (UniqueName: \"kubernetes.io/projected/edf3373e-6c91-419a-91b8-3cf9cd9e9725-kube-api-access-5znm9\") pod \"collect-profiles-29398050-hghk6\" (UID: \"edf3373e-6c91-419a-91b8-3cf9cd9e9725\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-hghk6" Nov 23 07:30:00 crc kubenswrapper[4559]: I1123 07:30:00.291000 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/edf3373e-6c91-419a-91b8-3cf9cd9e9725-config-volume\") pod \"collect-profiles-29398050-hghk6\" (UID: \"edf3373e-6c91-419a-91b8-3cf9cd9e9725\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-hghk6" Nov 23 07:30:00 crc kubenswrapper[4559]: I1123 07:30:00.291092 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/edf3373e-6c91-419a-91b8-3cf9cd9e9725-secret-volume\") pod \"collect-profiles-29398050-hghk6\" (UID: \"edf3373e-6c91-419a-91b8-3cf9cd9e9725\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-hghk6" Nov 23 07:30:00 crc kubenswrapper[4559]: I1123 07:30:00.291832 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/edf3373e-6c91-419a-91b8-3cf9cd9e9725-config-volume\") pod \"collect-profiles-29398050-hghk6\" (UID: \"edf3373e-6c91-419a-91b8-3cf9cd9e9725\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-hghk6" Nov 23 07:30:00 crc kubenswrapper[4559]: I1123 07:30:00.295886 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/edf3373e-6c91-419a-91b8-3cf9cd9e9725-secret-volume\") pod \"collect-profiles-29398050-hghk6\" (UID: \"edf3373e-6c91-419a-91b8-3cf9cd9e9725\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-hghk6" Nov 23 07:30:00 crc kubenswrapper[4559]: I1123 07:30:00.303490 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5znm9\" (UniqueName: \"kubernetes.io/projected/edf3373e-6c91-419a-91b8-3cf9cd9e9725-kube-api-access-5znm9\") pod \"collect-profiles-29398050-hghk6\" (UID: \"edf3373e-6c91-419a-91b8-3cf9cd9e9725\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-hghk6" Nov 23 07:30:00 crc kubenswrapper[4559]: I1123 07:30:00.457525 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-hghk6" Nov 23 07:30:00 crc kubenswrapper[4559]: I1123 07:30:00.839918 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398050-hghk6"] Nov 23 07:30:03 crc kubenswrapper[4559]: W1123 07:30:03.702996 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podedf3373e_6c91_419a_91b8_3cf9cd9e9725.slice/crio-fea6b31317e42ebe810abd0a02408c183b8bb064e9fe521bc0d46f5902a2cb69 WatchSource:0}: Error finding container fea6b31317e42ebe810abd0a02408c183b8bb064e9fe521bc0d46f5902a2cb69: Status 404 returned error can't find the container with id fea6b31317e42ebe810abd0a02408c183b8bb064e9fe521bc0d46f5902a2cb69 Nov 23 07:30:03 crc kubenswrapper[4559]: I1123 07:30:03.852050 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xdv92"] Nov 23 07:30:03 crc kubenswrapper[4559]: I1123 07:30:03.854830 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xdv92" Nov 23 07:30:03 crc kubenswrapper[4559]: I1123 07:30:03.859050 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xdv92"] Nov 23 07:30:03 crc kubenswrapper[4559]: I1123 07:30:03.973839 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20a5cc1d-7883-4a80-bebc-32a05e6af0aa-utilities\") pod \"certified-operators-xdv92\" (UID: \"20a5cc1d-7883-4a80-bebc-32a05e6af0aa\") " pod="openshift-marketplace/certified-operators-xdv92" Nov 23 07:30:03 crc kubenswrapper[4559]: I1123 07:30:03.973913 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xshm\" (UniqueName: \"kubernetes.io/projected/20a5cc1d-7883-4a80-bebc-32a05e6af0aa-kube-api-access-8xshm\") pod \"certified-operators-xdv92\" (UID: \"20a5cc1d-7883-4a80-bebc-32a05e6af0aa\") " pod="openshift-marketplace/certified-operators-xdv92" Nov 23 07:30:03 crc kubenswrapper[4559]: I1123 07:30:03.973935 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20a5cc1d-7883-4a80-bebc-32a05e6af0aa-catalog-content\") pod \"certified-operators-xdv92\" (UID: \"20a5cc1d-7883-4a80-bebc-32a05e6af0aa\") " pod="openshift-marketplace/certified-operators-xdv92" Nov 23 07:30:04 crc kubenswrapper[4559]: I1123 07:30:04.052716 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-hghk6" event={"ID":"edf3373e-6c91-419a-91b8-3cf9cd9e9725","Type":"ContainerStarted","Data":"56dc392dccab0b45e106278c94a0b30050bb6d707a3cc29d0ad8fbebe2accb9a"} Nov 23 07:30:04 crc kubenswrapper[4559]: I1123 07:30:04.052763 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-hghk6" event={"ID":"edf3373e-6c91-419a-91b8-3cf9cd9e9725","Type":"ContainerStarted","Data":"fea6b31317e42ebe810abd0a02408c183b8bb064e9fe521bc0d46f5902a2cb69"} Nov 23 07:30:04 crc kubenswrapper[4559]: I1123 07:30:04.054460 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m8wzp/must-gather-8nhbp" event={"ID":"13525b40-ea0a-4894-8622-e017b4f34397","Type":"ContainerStarted","Data":"956585f1ea41a294bbff57232074686727d455815232ba5506a96bcfde6a105a"} Nov 23 07:30:04 crc kubenswrapper[4559]: I1123 07:30:04.069878 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-hghk6" podStartSLOduration=4.069861684 podStartE2EDuration="4.069861684s" podCreationTimestamp="2025-11-23 07:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:30:04.064763469 +0000 UTC m=+2706.086749083" watchObservedRunningTime="2025-11-23 07:30:04.069861684 +0000 UTC m=+2706.091847297" Nov 23 07:30:04 crc kubenswrapper[4559]: I1123 07:30:04.076290 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20a5cc1d-7883-4a80-bebc-32a05e6af0aa-utilities\") pod \"certified-operators-xdv92\" (UID: \"20a5cc1d-7883-4a80-bebc-32a05e6af0aa\") " pod="openshift-marketplace/certified-operators-xdv92" Nov 23 07:30:04 crc kubenswrapper[4559]: I1123 07:30:04.076413 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xshm\" (UniqueName: \"kubernetes.io/projected/20a5cc1d-7883-4a80-bebc-32a05e6af0aa-kube-api-access-8xshm\") pod \"certified-operators-xdv92\" (UID: \"20a5cc1d-7883-4a80-bebc-32a05e6af0aa\") " pod="openshift-marketplace/certified-operators-xdv92" Nov 23 07:30:04 crc kubenswrapper[4559]: I1123 07:30:04.076440 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20a5cc1d-7883-4a80-bebc-32a05e6af0aa-catalog-content\") pod \"certified-operators-xdv92\" (UID: \"20a5cc1d-7883-4a80-bebc-32a05e6af0aa\") " pod="openshift-marketplace/certified-operators-xdv92" Nov 23 07:30:04 crc kubenswrapper[4559]: I1123 07:30:04.078666 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20a5cc1d-7883-4a80-bebc-32a05e6af0aa-utilities\") pod \"certified-operators-xdv92\" (UID: \"20a5cc1d-7883-4a80-bebc-32a05e6af0aa\") " pod="openshift-marketplace/certified-operators-xdv92" Nov 23 07:30:04 crc kubenswrapper[4559]: I1123 07:30:04.079162 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20a5cc1d-7883-4a80-bebc-32a05e6af0aa-catalog-content\") pod \"certified-operators-xdv92\" (UID: \"20a5cc1d-7883-4a80-bebc-32a05e6af0aa\") " pod="openshift-marketplace/certified-operators-xdv92" Nov 23 07:30:04 crc kubenswrapper[4559]: I1123 07:30:04.099381 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xshm\" (UniqueName: \"kubernetes.io/projected/20a5cc1d-7883-4a80-bebc-32a05e6af0aa-kube-api-access-8xshm\") pod \"certified-operators-xdv92\" (UID: \"20a5cc1d-7883-4a80-bebc-32a05e6af0aa\") " pod="openshift-marketplace/certified-operators-xdv92" Nov 23 07:30:04 crc kubenswrapper[4559]: I1123 07:30:04.183992 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xdv92" Nov 23 07:30:04 crc kubenswrapper[4559]: W1123 07:30:04.669167 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20a5cc1d_7883_4a80_bebc_32a05e6af0aa.slice/crio-752512559adcd8b85900503d73b6c76347f5f5b4246d138ecda9408fcfd12b7e WatchSource:0}: Error finding container 752512559adcd8b85900503d73b6c76347f5f5b4246d138ecda9408fcfd12b7e: Status 404 returned error can't find the container with id 752512559adcd8b85900503d73b6c76347f5f5b4246d138ecda9408fcfd12b7e Nov 23 07:30:04 crc kubenswrapper[4559]: I1123 07:30:04.670346 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xdv92"] Nov 23 07:30:05 crc kubenswrapper[4559]: I1123 07:30:05.068831 4559 generic.go:334] "Generic (PLEG): container finished" podID="20a5cc1d-7883-4a80-bebc-32a05e6af0aa" containerID="34dc6894c999b5aa4c402296b3d4323f19c0994a18bfd476997a9d240ae6944e" exitCode=0 Nov 23 07:30:05 crc kubenswrapper[4559]: I1123 07:30:05.068909 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xdv92" event={"ID":"20a5cc1d-7883-4a80-bebc-32a05e6af0aa","Type":"ContainerDied","Data":"34dc6894c999b5aa4c402296b3d4323f19c0994a18bfd476997a9d240ae6944e"} Nov 23 07:30:05 crc kubenswrapper[4559]: I1123 07:30:05.068956 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xdv92" event={"ID":"20a5cc1d-7883-4a80-bebc-32a05e6af0aa","Type":"ContainerStarted","Data":"752512559adcd8b85900503d73b6c76347f5f5b4246d138ecda9408fcfd12b7e"} Nov 23 07:30:05 crc kubenswrapper[4559]: I1123 07:30:05.074061 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m8wzp/must-gather-8nhbp" event={"ID":"13525b40-ea0a-4894-8622-e017b4f34397","Type":"ContainerStarted","Data":"5499865a9a9dd4b62b7fc407abdf7e24b86e7fee32f9b16260a32c7058847885"} Nov 23 07:30:05 crc kubenswrapper[4559]: I1123 07:30:05.079720 4559 generic.go:334] "Generic (PLEG): container finished" podID="edf3373e-6c91-419a-91b8-3cf9cd9e9725" containerID="56dc392dccab0b45e106278c94a0b30050bb6d707a3cc29d0ad8fbebe2accb9a" exitCode=0 Nov 23 07:30:05 crc kubenswrapper[4559]: I1123 07:30:05.079764 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-hghk6" event={"ID":"edf3373e-6c91-419a-91b8-3cf9cd9e9725","Type":"ContainerDied","Data":"56dc392dccab0b45e106278c94a0b30050bb6d707a3cc29d0ad8fbebe2accb9a"} Nov 23 07:30:05 crc kubenswrapper[4559]: I1123 07:30:05.143119 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-m8wzp/must-gather-8nhbp" podStartSLOduration=2.306249326 podStartE2EDuration="7.143102052s" podCreationTimestamp="2025-11-23 07:29:58 +0000 UTC" firstStartedPulling="2025-11-23 07:29:58.9003125 +0000 UTC m=+2700.922298114" lastFinishedPulling="2025-11-23 07:30:03.737165226 +0000 UTC m=+2705.759150840" observedRunningTime="2025-11-23 07:30:05.138912325 +0000 UTC m=+2707.160897940" watchObservedRunningTime="2025-11-23 07:30:05.143102052 +0000 UTC m=+2707.165087665" Nov 23 07:30:06 crc kubenswrapper[4559]: I1123 07:30:06.095442 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xdv92" event={"ID":"20a5cc1d-7883-4a80-bebc-32a05e6af0aa","Type":"ContainerStarted","Data":"531c177f2e5813f458f29a124581d9a8e099a9f7be0f8c304740fc119d48c6d5"} Nov 23 07:30:06 crc kubenswrapper[4559]: I1123 07:30:06.391338 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-hghk6" Nov 23 07:30:06 crc kubenswrapper[4559]: I1123 07:30:06.430453 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5znm9\" (UniqueName: \"kubernetes.io/projected/edf3373e-6c91-419a-91b8-3cf9cd9e9725-kube-api-access-5znm9\") pod \"edf3373e-6c91-419a-91b8-3cf9cd9e9725\" (UID: \"edf3373e-6c91-419a-91b8-3cf9cd9e9725\") " Nov 23 07:30:06 crc kubenswrapper[4559]: I1123 07:30:06.430563 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/edf3373e-6c91-419a-91b8-3cf9cd9e9725-config-volume\") pod \"edf3373e-6c91-419a-91b8-3cf9cd9e9725\" (UID: \"edf3373e-6c91-419a-91b8-3cf9cd9e9725\") " Nov 23 07:30:06 crc kubenswrapper[4559]: I1123 07:30:06.430833 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/edf3373e-6c91-419a-91b8-3cf9cd9e9725-secret-volume\") pod \"edf3373e-6c91-419a-91b8-3cf9cd9e9725\" (UID: \"edf3373e-6c91-419a-91b8-3cf9cd9e9725\") " Nov 23 07:30:06 crc kubenswrapper[4559]: I1123 07:30:06.432843 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edf3373e-6c91-419a-91b8-3cf9cd9e9725-config-volume" (OuterVolumeSpecName: "config-volume") pod "edf3373e-6c91-419a-91b8-3cf9cd9e9725" (UID: "edf3373e-6c91-419a-91b8-3cf9cd9e9725"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 07:30:06 crc kubenswrapper[4559]: I1123 07:30:06.438790 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edf3373e-6c91-419a-91b8-3cf9cd9e9725-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "edf3373e-6c91-419a-91b8-3cf9cd9e9725" (UID: "edf3373e-6c91-419a-91b8-3cf9cd9e9725"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 07:30:06 crc kubenswrapper[4559]: I1123 07:30:06.439755 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edf3373e-6c91-419a-91b8-3cf9cd9e9725-kube-api-access-5znm9" (OuterVolumeSpecName: "kube-api-access-5znm9") pod "edf3373e-6c91-419a-91b8-3cf9cd9e9725" (UID: "edf3373e-6c91-419a-91b8-3cf9cd9e9725"). InnerVolumeSpecName "kube-api-access-5znm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:30:06 crc kubenswrapper[4559]: I1123 07:30:06.534116 4559 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/edf3373e-6c91-419a-91b8-3cf9cd9e9725-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 23 07:30:06 crc kubenswrapper[4559]: I1123 07:30:06.534156 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5znm9\" (UniqueName: \"kubernetes.io/projected/edf3373e-6c91-419a-91b8-3cf9cd9e9725-kube-api-access-5znm9\") on node \"crc\" DevicePath \"\"" Nov 23 07:30:06 crc kubenswrapper[4559]: I1123 07:30:06.534166 4559 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/edf3373e-6c91-419a-91b8-3cf9cd9e9725-config-volume\") on node \"crc\" DevicePath \"\"" Nov 23 07:30:07 crc kubenswrapper[4559]: I1123 07:30:07.044733 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-m8wzp/crc-debug-r6lhl"] Nov 23 07:30:07 crc kubenswrapper[4559]: E1123 07:30:07.045084 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edf3373e-6c91-419a-91b8-3cf9cd9e9725" containerName="collect-profiles" Nov 23 07:30:07 crc kubenswrapper[4559]: I1123 07:30:07.045101 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="edf3373e-6c91-419a-91b8-3cf9cd9e9725" containerName="collect-profiles" Nov 23 07:30:07 crc kubenswrapper[4559]: I1123 07:30:07.045284 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="edf3373e-6c91-419a-91b8-3cf9cd9e9725" containerName="collect-profiles" Nov 23 07:30:07 crc kubenswrapper[4559]: I1123 07:30:07.045853 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m8wzp/crc-debug-r6lhl" Nov 23 07:30:07 crc kubenswrapper[4559]: I1123 07:30:07.104555 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-hghk6" Nov 23 07:30:07 crc kubenswrapper[4559]: I1123 07:30:07.104538 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398050-hghk6" event={"ID":"edf3373e-6c91-419a-91b8-3cf9cd9e9725","Type":"ContainerDied","Data":"fea6b31317e42ebe810abd0a02408c183b8bb064e9fe521bc0d46f5902a2cb69"} Nov 23 07:30:07 crc kubenswrapper[4559]: I1123 07:30:07.104679 4559 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fea6b31317e42ebe810abd0a02408c183b8bb064e9fe521bc0d46f5902a2cb69" Nov 23 07:30:07 crc kubenswrapper[4559]: I1123 07:30:07.106485 4559 generic.go:334] "Generic (PLEG): container finished" podID="20a5cc1d-7883-4a80-bebc-32a05e6af0aa" containerID="531c177f2e5813f458f29a124581d9a8e099a9f7be0f8c304740fc119d48c6d5" exitCode=0 Nov 23 07:30:07 crc kubenswrapper[4559]: I1123 07:30:07.106515 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xdv92" event={"ID":"20a5cc1d-7883-4a80-bebc-32a05e6af0aa","Type":"ContainerDied","Data":"531c177f2e5813f458f29a124581d9a8e099a9f7be0f8c304740fc119d48c6d5"} Nov 23 07:30:07 crc kubenswrapper[4559]: I1123 07:30:07.143308 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398005-nnldq"] Nov 23 07:30:07 crc kubenswrapper[4559]: I1123 07:30:07.144391 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwz8r\" (UniqueName: \"kubernetes.io/projected/9a8dca5e-a751-49d9-94a7-7a91af8117f0-kube-api-access-bwz8r\") pod \"crc-debug-r6lhl\" (UID: \"9a8dca5e-a751-49d9-94a7-7a91af8117f0\") " pod="openshift-must-gather-m8wzp/crc-debug-r6lhl" Nov 23 07:30:07 crc kubenswrapper[4559]: I1123 07:30:07.144677 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9a8dca5e-a751-49d9-94a7-7a91af8117f0-host\") pod \"crc-debug-r6lhl\" (UID: \"9a8dca5e-a751-49d9-94a7-7a91af8117f0\") " pod="openshift-must-gather-m8wzp/crc-debug-r6lhl" Nov 23 07:30:07 crc kubenswrapper[4559]: I1123 07:30:07.149807 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398005-nnldq"] Nov 23 07:30:07 crc kubenswrapper[4559]: I1123 07:30:07.245960 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9a8dca5e-a751-49d9-94a7-7a91af8117f0-host\") pod \"crc-debug-r6lhl\" (UID: \"9a8dca5e-a751-49d9-94a7-7a91af8117f0\") " pod="openshift-must-gather-m8wzp/crc-debug-r6lhl" Nov 23 07:30:07 crc kubenswrapper[4559]: I1123 07:30:07.246081 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9a8dca5e-a751-49d9-94a7-7a91af8117f0-host\") pod \"crc-debug-r6lhl\" (UID: \"9a8dca5e-a751-49d9-94a7-7a91af8117f0\") " pod="openshift-must-gather-m8wzp/crc-debug-r6lhl" Nov 23 07:30:07 crc kubenswrapper[4559]: I1123 07:30:07.246191 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwz8r\" (UniqueName: \"kubernetes.io/projected/9a8dca5e-a751-49d9-94a7-7a91af8117f0-kube-api-access-bwz8r\") pod \"crc-debug-r6lhl\" (UID: \"9a8dca5e-a751-49d9-94a7-7a91af8117f0\") " pod="openshift-must-gather-m8wzp/crc-debug-r6lhl" Nov 23 07:30:07 crc kubenswrapper[4559]: I1123 07:30:07.266303 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwz8r\" (UniqueName: \"kubernetes.io/projected/9a8dca5e-a751-49d9-94a7-7a91af8117f0-kube-api-access-bwz8r\") pod \"crc-debug-r6lhl\" (UID: \"9a8dca5e-a751-49d9-94a7-7a91af8117f0\") " pod="openshift-must-gather-m8wzp/crc-debug-r6lhl" Nov 23 07:30:07 crc kubenswrapper[4559]: I1123 07:30:07.361316 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m8wzp/crc-debug-r6lhl" Nov 23 07:30:08 crc kubenswrapper[4559]: I1123 07:30:08.115489 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xdv92" event={"ID":"20a5cc1d-7883-4a80-bebc-32a05e6af0aa","Type":"ContainerStarted","Data":"68617a6916f527baaae33d51553663924f1ad10290b240a05fe11d2b5cdce3c6"} Nov 23 07:30:08 crc kubenswrapper[4559]: I1123 07:30:08.116699 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m8wzp/crc-debug-r6lhl" event={"ID":"9a8dca5e-a751-49d9-94a7-7a91af8117f0","Type":"ContainerStarted","Data":"ed8b22e6f87bc6333754be173f12f0f0225dd16f77539ead9d167d7e36a76e6f"} Nov 23 07:30:08 crc kubenswrapper[4559]: I1123 07:30:08.133717 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xdv92" podStartSLOduration=2.617753584 podStartE2EDuration="5.133702359s" podCreationTimestamp="2025-11-23 07:30:03 +0000 UTC" firstStartedPulling="2025-11-23 07:30:05.070564351 +0000 UTC m=+2707.092549965" lastFinishedPulling="2025-11-23 07:30:07.586513126 +0000 UTC m=+2709.608498740" observedRunningTime="2025-11-23 07:30:08.127749488 +0000 UTC m=+2710.149735102" watchObservedRunningTime="2025-11-23 07:30:08.133702359 +0000 UTC m=+2710.155687973" Nov 23 07:30:08 crc kubenswrapper[4559]: I1123 07:30:08.282033 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1d0d430-031f-448e-9763-69deb17abc4a" path="/var/lib/kubelet/pods/d1d0d430-031f-448e-9763-69deb17abc4a/volumes" Nov 23 07:30:14 crc kubenswrapper[4559]: I1123 07:30:14.184571 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xdv92" Nov 23 07:30:14 crc kubenswrapper[4559]: I1123 07:30:14.184869 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xdv92" Nov 23 07:30:14 crc kubenswrapper[4559]: I1123 07:30:14.227908 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xdv92" Nov 23 07:30:15 crc kubenswrapper[4559]: I1123 07:30:15.143351 4559 scope.go:117] "RemoveContainer" containerID="fb8c83c3dc960f919ee7c61b2b0f2954d8b86e86288f9ed98a5ac3db1f7fdc4e" Nov 23 07:30:15 crc kubenswrapper[4559]: I1123 07:30:15.202018 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xdv92" Nov 23 07:30:15 crc kubenswrapper[4559]: I1123 07:30:15.253723 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xdv92"] Nov 23 07:30:17 crc kubenswrapper[4559]: I1123 07:30:17.184635 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m8wzp/crc-debug-r6lhl" event={"ID":"9a8dca5e-a751-49d9-94a7-7a91af8117f0","Type":"ContainerStarted","Data":"846f39825d6c99fab5b232a71bb114afaa83f91ae82c33b8f2e6a60887341ffe"} Nov 23 07:30:17 crc kubenswrapper[4559]: I1123 07:30:17.184790 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xdv92" podUID="20a5cc1d-7883-4a80-bebc-32a05e6af0aa" containerName="registry-server" containerID="cri-o://68617a6916f527baaae33d51553663924f1ad10290b240a05fe11d2b5cdce3c6" gracePeriod=2 Nov 23 07:30:17 crc kubenswrapper[4559]: I1123 07:30:17.203483 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-m8wzp/crc-debug-r6lhl" podStartSLOduration=0.877757473 podStartE2EDuration="10.203468066s" podCreationTimestamp="2025-11-23 07:30:07 +0000 UTC" firstStartedPulling="2025-11-23 07:30:07.386534646 +0000 UTC m=+2709.408520261" lastFinishedPulling="2025-11-23 07:30:16.71224524 +0000 UTC m=+2718.734230854" observedRunningTime="2025-11-23 07:30:17.196420427 +0000 UTC m=+2719.218406041" watchObservedRunningTime="2025-11-23 07:30:17.203468066 +0000 UTC m=+2719.225453681" Nov 23 07:30:17 crc kubenswrapper[4559]: I1123 07:30:17.549763 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xdv92" Nov 23 07:30:17 crc kubenswrapper[4559]: I1123 07:30:17.729817 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20a5cc1d-7883-4a80-bebc-32a05e6af0aa-utilities\") pod \"20a5cc1d-7883-4a80-bebc-32a05e6af0aa\" (UID: \"20a5cc1d-7883-4a80-bebc-32a05e6af0aa\") " Nov 23 07:30:17 crc kubenswrapper[4559]: I1123 07:30:17.729977 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xshm\" (UniqueName: \"kubernetes.io/projected/20a5cc1d-7883-4a80-bebc-32a05e6af0aa-kube-api-access-8xshm\") pod \"20a5cc1d-7883-4a80-bebc-32a05e6af0aa\" (UID: \"20a5cc1d-7883-4a80-bebc-32a05e6af0aa\") " Nov 23 07:30:17 crc kubenswrapper[4559]: I1123 07:30:17.729999 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20a5cc1d-7883-4a80-bebc-32a05e6af0aa-catalog-content\") pod \"20a5cc1d-7883-4a80-bebc-32a05e6af0aa\" (UID: \"20a5cc1d-7883-4a80-bebc-32a05e6af0aa\") " Nov 23 07:30:17 crc kubenswrapper[4559]: I1123 07:30:17.730313 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20a5cc1d-7883-4a80-bebc-32a05e6af0aa-utilities" (OuterVolumeSpecName: "utilities") pod "20a5cc1d-7883-4a80-bebc-32a05e6af0aa" (UID: "20a5cc1d-7883-4a80-bebc-32a05e6af0aa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:30:17 crc kubenswrapper[4559]: I1123 07:30:17.730787 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20a5cc1d-7883-4a80-bebc-32a05e6af0aa-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:30:17 crc kubenswrapper[4559]: I1123 07:30:17.742139 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20a5cc1d-7883-4a80-bebc-32a05e6af0aa-kube-api-access-8xshm" (OuterVolumeSpecName: "kube-api-access-8xshm") pod "20a5cc1d-7883-4a80-bebc-32a05e6af0aa" (UID: "20a5cc1d-7883-4a80-bebc-32a05e6af0aa"). InnerVolumeSpecName "kube-api-access-8xshm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:30:17 crc kubenswrapper[4559]: I1123 07:30:17.764946 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20a5cc1d-7883-4a80-bebc-32a05e6af0aa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "20a5cc1d-7883-4a80-bebc-32a05e6af0aa" (UID: "20a5cc1d-7883-4a80-bebc-32a05e6af0aa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:30:17 crc kubenswrapper[4559]: I1123 07:30:17.832061 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xshm\" (UniqueName: \"kubernetes.io/projected/20a5cc1d-7883-4a80-bebc-32a05e6af0aa-kube-api-access-8xshm\") on node \"crc\" DevicePath \"\"" Nov 23 07:30:17 crc kubenswrapper[4559]: I1123 07:30:17.832315 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20a5cc1d-7883-4a80-bebc-32a05e6af0aa-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:30:18 crc kubenswrapper[4559]: I1123 07:30:18.196170 4559 generic.go:334] "Generic (PLEG): container finished" podID="20a5cc1d-7883-4a80-bebc-32a05e6af0aa" containerID="68617a6916f527baaae33d51553663924f1ad10290b240a05fe11d2b5cdce3c6" exitCode=0 Nov 23 07:30:18 crc kubenswrapper[4559]: I1123 07:30:18.197016 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xdv92" Nov 23 07:30:18 crc kubenswrapper[4559]: I1123 07:30:18.204296 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xdv92" event={"ID":"20a5cc1d-7883-4a80-bebc-32a05e6af0aa","Type":"ContainerDied","Data":"68617a6916f527baaae33d51553663924f1ad10290b240a05fe11d2b5cdce3c6"} Nov 23 07:30:18 crc kubenswrapper[4559]: I1123 07:30:18.204324 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xdv92" event={"ID":"20a5cc1d-7883-4a80-bebc-32a05e6af0aa","Type":"ContainerDied","Data":"752512559adcd8b85900503d73b6c76347f5f5b4246d138ecda9408fcfd12b7e"} Nov 23 07:30:18 crc kubenswrapper[4559]: I1123 07:30:18.204341 4559 scope.go:117] "RemoveContainer" containerID="68617a6916f527baaae33d51553663924f1ad10290b240a05fe11d2b5cdce3c6" Nov 23 07:30:18 crc kubenswrapper[4559]: I1123 07:30:18.242572 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xdv92"] Nov 23 07:30:18 crc kubenswrapper[4559]: I1123 07:30:18.248412 4559 scope.go:117] "RemoveContainer" containerID="531c177f2e5813f458f29a124581d9a8e099a9f7be0f8c304740fc119d48c6d5" Nov 23 07:30:18 crc kubenswrapper[4559]: I1123 07:30:18.252569 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xdv92"] Nov 23 07:30:18 crc kubenswrapper[4559]: I1123 07:30:18.272390 4559 scope.go:117] "RemoveContainer" containerID="34dc6894c999b5aa4c402296b3d4323f19c0994a18bfd476997a9d240ae6944e" Nov 23 07:30:18 crc kubenswrapper[4559]: I1123 07:30:18.284120 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20a5cc1d-7883-4a80-bebc-32a05e6af0aa" path="/var/lib/kubelet/pods/20a5cc1d-7883-4a80-bebc-32a05e6af0aa/volumes" Nov 23 07:30:18 crc kubenswrapper[4559]: I1123 07:30:18.304775 4559 scope.go:117] "RemoveContainer" containerID="68617a6916f527baaae33d51553663924f1ad10290b240a05fe11d2b5cdce3c6" Nov 23 07:30:18 crc kubenswrapper[4559]: E1123 07:30:18.305804 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68617a6916f527baaae33d51553663924f1ad10290b240a05fe11d2b5cdce3c6\": container with ID starting with 68617a6916f527baaae33d51553663924f1ad10290b240a05fe11d2b5cdce3c6 not found: ID does not exist" containerID="68617a6916f527baaae33d51553663924f1ad10290b240a05fe11d2b5cdce3c6" Nov 23 07:30:18 crc kubenswrapper[4559]: I1123 07:30:18.305839 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68617a6916f527baaae33d51553663924f1ad10290b240a05fe11d2b5cdce3c6"} err="failed to get container status \"68617a6916f527baaae33d51553663924f1ad10290b240a05fe11d2b5cdce3c6\": rpc error: code = NotFound desc = could not find container \"68617a6916f527baaae33d51553663924f1ad10290b240a05fe11d2b5cdce3c6\": container with ID starting with 68617a6916f527baaae33d51553663924f1ad10290b240a05fe11d2b5cdce3c6 not found: ID does not exist" Nov 23 07:30:18 crc kubenswrapper[4559]: I1123 07:30:18.305875 4559 scope.go:117] "RemoveContainer" containerID="531c177f2e5813f458f29a124581d9a8e099a9f7be0f8c304740fc119d48c6d5" Nov 23 07:30:18 crc kubenswrapper[4559]: E1123 07:30:18.307209 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"531c177f2e5813f458f29a124581d9a8e099a9f7be0f8c304740fc119d48c6d5\": container with ID starting with 531c177f2e5813f458f29a124581d9a8e099a9f7be0f8c304740fc119d48c6d5 not found: ID does not exist" containerID="531c177f2e5813f458f29a124581d9a8e099a9f7be0f8c304740fc119d48c6d5" Nov 23 07:30:18 crc kubenswrapper[4559]: I1123 07:30:18.307248 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"531c177f2e5813f458f29a124581d9a8e099a9f7be0f8c304740fc119d48c6d5"} err="failed to get container status \"531c177f2e5813f458f29a124581d9a8e099a9f7be0f8c304740fc119d48c6d5\": rpc error: code = NotFound desc = could not find container \"531c177f2e5813f458f29a124581d9a8e099a9f7be0f8c304740fc119d48c6d5\": container with ID starting with 531c177f2e5813f458f29a124581d9a8e099a9f7be0f8c304740fc119d48c6d5 not found: ID does not exist" Nov 23 07:30:18 crc kubenswrapper[4559]: I1123 07:30:18.307276 4559 scope.go:117] "RemoveContainer" containerID="34dc6894c999b5aa4c402296b3d4323f19c0994a18bfd476997a9d240ae6944e" Nov 23 07:30:18 crc kubenswrapper[4559]: E1123 07:30:18.307740 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34dc6894c999b5aa4c402296b3d4323f19c0994a18bfd476997a9d240ae6944e\": container with ID starting with 34dc6894c999b5aa4c402296b3d4323f19c0994a18bfd476997a9d240ae6944e not found: ID does not exist" containerID="34dc6894c999b5aa4c402296b3d4323f19c0994a18bfd476997a9d240ae6944e" Nov 23 07:30:18 crc kubenswrapper[4559]: I1123 07:30:18.307782 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34dc6894c999b5aa4c402296b3d4323f19c0994a18bfd476997a9d240ae6944e"} err="failed to get container status \"34dc6894c999b5aa4c402296b3d4323f19c0994a18bfd476997a9d240ae6944e\": rpc error: code = NotFound desc = could not find container \"34dc6894c999b5aa4c402296b3d4323f19c0994a18bfd476997a9d240ae6944e\": container with ID starting with 34dc6894c999b5aa4c402296b3d4323f19c0994a18bfd476997a9d240ae6944e not found: ID does not exist" Nov 23 07:30:50 crc kubenswrapper[4559]: I1123 07:30:50.149123 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nw54h"] Nov 23 07:30:50 crc kubenswrapper[4559]: E1123 07:30:50.149800 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20a5cc1d-7883-4a80-bebc-32a05e6af0aa" containerName="extract-utilities" Nov 23 07:30:50 crc kubenswrapper[4559]: I1123 07:30:50.149813 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="20a5cc1d-7883-4a80-bebc-32a05e6af0aa" containerName="extract-utilities" Nov 23 07:30:50 crc kubenswrapper[4559]: E1123 07:30:50.149834 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20a5cc1d-7883-4a80-bebc-32a05e6af0aa" containerName="registry-server" Nov 23 07:30:50 crc kubenswrapper[4559]: I1123 07:30:50.149839 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="20a5cc1d-7883-4a80-bebc-32a05e6af0aa" containerName="registry-server" Nov 23 07:30:50 crc kubenswrapper[4559]: E1123 07:30:50.149857 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20a5cc1d-7883-4a80-bebc-32a05e6af0aa" containerName="extract-content" Nov 23 07:30:50 crc kubenswrapper[4559]: I1123 07:30:50.149862 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="20a5cc1d-7883-4a80-bebc-32a05e6af0aa" containerName="extract-content" Nov 23 07:30:50 crc kubenswrapper[4559]: I1123 07:30:50.150027 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="20a5cc1d-7883-4a80-bebc-32a05e6af0aa" containerName="registry-server" Nov 23 07:30:50 crc kubenswrapper[4559]: I1123 07:30:50.151235 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nw54h" Nov 23 07:30:50 crc kubenswrapper[4559]: I1123 07:30:50.160083 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nw54h"] Nov 23 07:30:50 crc kubenswrapper[4559]: I1123 07:30:50.262496 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77tqw\" (UniqueName: \"kubernetes.io/projected/c2daeb5e-241f-4764-962b-64c9c3d34597-kube-api-access-77tqw\") pod \"community-operators-nw54h\" (UID: \"c2daeb5e-241f-4764-962b-64c9c3d34597\") " pod="openshift-marketplace/community-operators-nw54h" Nov 23 07:30:50 crc kubenswrapper[4559]: I1123 07:30:50.262608 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2daeb5e-241f-4764-962b-64c9c3d34597-catalog-content\") pod \"community-operators-nw54h\" (UID: \"c2daeb5e-241f-4764-962b-64c9c3d34597\") " pod="openshift-marketplace/community-operators-nw54h" Nov 23 07:30:50 crc kubenswrapper[4559]: I1123 07:30:50.262680 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2daeb5e-241f-4764-962b-64c9c3d34597-utilities\") pod \"community-operators-nw54h\" (UID: \"c2daeb5e-241f-4764-962b-64c9c3d34597\") " pod="openshift-marketplace/community-operators-nw54h" Nov 23 07:30:50 crc kubenswrapper[4559]: I1123 07:30:50.364366 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2daeb5e-241f-4764-962b-64c9c3d34597-catalog-content\") pod \"community-operators-nw54h\" (UID: \"c2daeb5e-241f-4764-962b-64c9c3d34597\") " pod="openshift-marketplace/community-operators-nw54h" Nov 23 07:30:50 crc kubenswrapper[4559]: I1123 07:30:50.364434 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2daeb5e-241f-4764-962b-64c9c3d34597-utilities\") pod \"community-operators-nw54h\" (UID: \"c2daeb5e-241f-4764-962b-64c9c3d34597\") " pod="openshift-marketplace/community-operators-nw54h" Nov 23 07:30:50 crc kubenswrapper[4559]: I1123 07:30:50.364586 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77tqw\" (UniqueName: \"kubernetes.io/projected/c2daeb5e-241f-4764-962b-64c9c3d34597-kube-api-access-77tqw\") pod \"community-operators-nw54h\" (UID: \"c2daeb5e-241f-4764-962b-64c9c3d34597\") " pod="openshift-marketplace/community-operators-nw54h" Nov 23 07:30:50 crc kubenswrapper[4559]: I1123 07:30:50.364891 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2daeb5e-241f-4764-962b-64c9c3d34597-catalog-content\") pod \"community-operators-nw54h\" (UID: \"c2daeb5e-241f-4764-962b-64c9c3d34597\") " pod="openshift-marketplace/community-operators-nw54h" Nov 23 07:30:50 crc kubenswrapper[4559]: I1123 07:30:50.365182 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2daeb5e-241f-4764-962b-64c9c3d34597-utilities\") pod \"community-operators-nw54h\" (UID: \"c2daeb5e-241f-4764-962b-64c9c3d34597\") " pod="openshift-marketplace/community-operators-nw54h" Nov 23 07:30:50 crc kubenswrapper[4559]: I1123 07:30:50.380819 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77tqw\" (UniqueName: \"kubernetes.io/projected/c2daeb5e-241f-4764-962b-64c9c3d34597-kube-api-access-77tqw\") pod \"community-operators-nw54h\" (UID: \"c2daeb5e-241f-4764-962b-64c9c3d34597\") " pod="openshift-marketplace/community-operators-nw54h" Nov 23 07:30:50 crc kubenswrapper[4559]: I1123 07:30:50.424977 4559 generic.go:334] "Generic (PLEG): container finished" podID="9a8dca5e-a751-49d9-94a7-7a91af8117f0" containerID="846f39825d6c99fab5b232a71bb114afaa83f91ae82c33b8f2e6a60887341ffe" exitCode=0 Nov 23 07:30:50 crc kubenswrapper[4559]: I1123 07:30:50.425044 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m8wzp/crc-debug-r6lhl" event={"ID":"9a8dca5e-a751-49d9-94a7-7a91af8117f0","Type":"ContainerDied","Data":"846f39825d6c99fab5b232a71bb114afaa83f91ae82c33b8f2e6a60887341ffe"} Nov 23 07:30:50 crc kubenswrapper[4559]: I1123 07:30:50.469822 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nw54h" Nov 23 07:30:50 crc kubenswrapper[4559]: I1123 07:30:50.860168 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nw54h"] Nov 23 07:30:51 crc kubenswrapper[4559]: I1123 07:30:51.432692 4559 generic.go:334] "Generic (PLEG): container finished" podID="c2daeb5e-241f-4764-962b-64c9c3d34597" containerID="83bd431bfc91f4c37e23d3636fe8a91f0793f37252ed4e644e2fe7c6bc368438" exitCode=0 Nov 23 07:30:51 crc kubenswrapper[4559]: I1123 07:30:51.433191 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nw54h" event={"ID":"c2daeb5e-241f-4764-962b-64c9c3d34597","Type":"ContainerDied","Data":"83bd431bfc91f4c37e23d3636fe8a91f0793f37252ed4e644e2fe7c6bc368438"} Nov 23 07:30:51 crc kubenswrapper[4559]: I1123 07:30:51.433232 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nw54h" event={"ID":"c2daeb5e-241f-4764-962b-64c9c3d34597","Type":"ContainerStarted","Data":"894c4303b79306e9a11a2f6d1f743caddaae2c251ba1efd320f8365e196d3785"} Nov 23 07:30:51 crc kubenswrapper[4559]: I1123 07:30:51.436126 4559 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 07:30:51 crc kubenswrapper[4559]: I1123 07:30:51.499460 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m8wzp/crc-debug-r6lhl" Nov 23 07:30:51 crc kubenswrapper[4559]: I1123 07:30:51.520350 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-m8wzp/crc-debug-r6lhl"] Nov 23 07:30:51 crc kubenswrapper[4559]: I1123 07:30:51.525988 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-m8wzp/crc-debug-r6lhl"] Nov 23 07:30:51 crc kubenswrapper[4559]: I1123 07:30:51.582485 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwz8r\" (UniqueName: \"kubernetes.io/projected/9a8dca5e-a751-49d9-94a7-7a91af8117f0-kube-api-access-bwz8r\") pod \"9a8dca5e-a751-49d9-94a7-7a91af8117f0\" (UID: \"9a8dca5e-a751-49d9-94a7-7a91af8117f0\") " Nov 23 07:30:51 crc kubenswrapper[4559]: I1123 07:30:51.582548 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9a8dca5e-a751-49d9-94a7-7a91af8117f0-host\") pod \"9a8dca5e-a751-49d9-94a7-7a91af8117f0\" (UID: \"9a8dca5e-a751-49d9-94a7-7a91af8117f0\") " Nov 23 07:30:51 crc kubenswrapper[4559]: I1123 07:30:51.582831 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a8dca5e-a751-49d9-94a7-7a91af8117f0-host" (OuterVolumeSpecName: "host") pod "9a8dca5e-a751-49d9-94a7-7a91af8117f0" (UID: "9a8dca5e-a751-49d9-94a7-7a91af8117f0"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:30:51 crc kubenswrapper[4559]: I1123 07:30:51.583091 4559 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9a8dca5e-a751-49d9-94a7-7a91af8117f0-host\") on node \"crc\" DevicePath \"\"" Nov 23 07:30:51 crc kubenswrapper[4559]: I1123 07:30:51.586854 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a8dca5e-a751-49d9-94a7-7a91af8117f0-kube-api-access-bwz8r" (OuterVolumeSpecName: "kube-api-access-bwz8r") pod "9a8dca5e-a751-49d9-94a7-7a91af8117f0" (UID: "9a8dca5e-a751-49d9-94a7-7a91af8117f0"). InnerVolumeSpecName "kube-api-access-bwz8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:30:51 crc kubenswrapper[4559]: I1123 07:30:51.684606 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwz8r\" (UniqueName: \"kubernetes.io/projected/9a8dca5e-a751-49d9-94a7-7a91af8117f0-kube-api-access-bwz8r\") on node \"crc\" DevicePath \"\"" Nov 23 07:30:52 crc kubenswrapper[4559]: I1123 07:30:52.281435 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a8dca5e-a751-49d9-94a7-7a91af8117f0" path="/var/lib/kubelet/pods/9a8dca5e-a751-49d9-94a7-7a91af8117f0/volumes" Nov 23 07:30:52 crc kubenswrapper[4559]: I1123 07:30:52.439708 4559 scope.go:117] "RemoveContainer" containerID="846f39825d6c99fab5b232a71bb114afaa83f91ae82c33b8f2e6a60887341ffe" Nov 23 07:30:52 crc kubenswrapper[4559]: I1123 07:30:52.439787 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m8wzp/crc-debug-r6lhl" Nov 23 07:30:52 crc kubenswrapper[4559]: I1123 07:30:52.633086 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-m8wzp/crc-debug-n625q"] Nov 23 07:30:52 crc kubenswrapper[4559]: E1123 07:30:52.633409 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a8dca5e-a751-49d9-94a7-7a91af8117f0" containerName="container-00" Nov 23 07:30:52 crc kubenswrapper[4559]: I1123 07:30:52.633425 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a8dca5e-a751-49d9-94a7-7a91af8117f0" containerName="container-00" Nov 23 07:30:52 crc kubenswrapper[4559]: I1123 07:30:52.633602 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a8dca5e-a751-49d9-94a7-7a91af8117f0" containerName="container-00" Nov 23 07:30:52 crc kubenswrapper[4559]: I1123 07:30:52.634117 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m8wzp/crc-debug-n625q" Nov 23 07:30:52 crc kubenswrapper[4559]: I1123 07:30:52.696892 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/be3a330c-2918-4c87-89ae-2565413a2f72-host\") pod \"crc-debug-n625q\" (UID: \"be3a330c-2918-4c87-89ae-2565413a2f72\") " pod="openshift-must-gather-m8wzp/crc-debug-n625q" Nov 23 07:30:52 crc kubenswrapper[4559]: I1123 07:30:52.697048 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gn94\" (UniqueName: \"kubernetes.io/projected/be3a330c-2918-4c87-89ae-2565413a2f72-kube-api-access-5gn94\") pod \"crc-debug-n625q\" (UID: \"be3a330c-2918-4c87-89ae-2565413a2f72\") " pod="openshift-must-gather-m8wzp/crc-debug-n625q" Nov 23 07:30:52 crc kubenswrapper[4559]: I1123 07:30:52.798861 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gn94\" (UniqueName: \"kubernetes.io/projected/be3a330c-2918-4c87-89ae-2565413a2f72-kube-api-access-5gn94\") pod \"crc-debug-n625q\" (UID: \"be3a330c-2918-4c87-89ae-2565413a2f72\") " pod="openshift-must-gather-m8wzp/crc-debug-n625q" Nov 23 07:30:52 crc kubenswrapper[4559]: I1123 07:30:52.798933 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/be3a330c-2918-4c87-89ae-2565413a2f72-host\") pod \"crc-debug-n625q\" (UID: \"be3a330c-2918-4c87-89ae-2565413a2f72\") " pod="openshift-must-gather-m8wzp/crc-debug-n625q" Nov 23 07:30:52 crc kubenswrapper[4559]: I1123 07:30:52.799115 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/be3a330c-2918-4c87-89ae-2565413a2f72-host\") pod \"crc-debug-n625q\" (UID: \"be3a330c-2918-4c87-89ae-2565413a2f72\") " pod="openshift-must-gather-m8wzp/crc-debug-n625q" Nov 23 07:30:52 crc kubenswrapper[4559]: I1123 07:30:52.813222 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gn94\" (UniqueName: \"kubernetes.io/projected/be3a330c-2918-4c87-89ae-2565413a2f72-kube-api-access-5gn94\") pod \"crc-debug-n625q\" (UID: \"be3a330c-2918-4c87-89ae-2565413a2f72\") " pod="openshift-must-gather-m8wzp/crc-debug-n625q" Nov 23 07:30:52 crc kubenswrapper[4559]: I1123 07:30:52.955097 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m8wzp/crc-debug-n625q" Nov 23 07:30:52 crc kubenswrapper[4559]: W1123 07:30:52.974793 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe3a330c_2918_4c87_89ae_2565413a2f72.slice/crio-69b997aea6bdc3aa706d55d5d0257dd0dd8887b8d11c8f3d8ce2e0fc784f182e WatchSource:0}: Error finding container 69b997aea6bdc3aa706d55d5d0257dd0dd8887b8d11c8f3d8ce2e0fc784f182e: Status 404 returned error can't find the container with id 69b997aea6bdc3aa706d55d5d0257dd0dd8887b8d11c8f3d8ce2e0fc784f182e Nov 23 07:30:53 crc kubenswrapper[4559]: I1123 07:30:53.447426 4559 generic.go:334] "Generic (PLEG): container finished" podID="be3a330c-2918-4c87-89ae-2565413a2f72" containerID="553a9f1601de390f4e7885031dc7f1165e83459f56bac5616affd7f3aa19a499" exitCode=0 Nov 23 07:30:53 crc kubenswrapper[4559]: I1123 07:30:53.447512 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m8wzp/crc-debug-n625q" event={"ID":"be3a330c-2918-4c87-89ae-2565413a2f72","Type":"ContainerDied","Data":"553a9f1601de390f4e7885031dc7f1165e83459f56bac5616affd7f3aa19a499"} Nov 23 07:30:53 crc kubenswrapper[4559]: I1123 07:30:53.447632 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m8wzp/crc-debug-n625q" event={"ID":"be3a330c-2918-4c87-89ae-2565413a2f72","Type":"ContainerStarted","Data":"69b997aea6bdc3aa706d55d5d0257dd0dd8887b8d11c8f3d8ce2e0fc784f182e"} Nov 23 07:30:53 crc kubenswrapper[4559]: I1123 07:30:53.925799 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-m8wzp/crc-debug-n625q"] Nov 23 07:30:53 crc kubenswrapper[4559]: I1123 07:30:53.932236 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-m8wzp/crc-debug-n625q"] Nov 23 07:30:55 crc kubenswrapper[4559]: I1123 07:30:55.041998 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-m8wzp/crc-debug-gv5q8"] Nov 23 07:30:55 crc kubenswrapper[4559]: E1123 07:30:55.042493 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be3a330c-2918-4c87-89ae-2565413a2f72" containerName="container-00" Nov 23 07:30:55 crc kubenswrapper[4559]: I1123 07:30:55.042505 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="be3a330c-2918-4c87-89ae-2565413a2f72" containerName="container-00" Nov 23 07:30:55 crc kubenswrapper[4559]: I1123 07:30:55.042708 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="be3a330c-2918-4c87-89ae-2565413a2f72" containerName="container-00" Nov 23 07:30:55 crc kubenswrapper[4559]: I1123 07:30:55.043191 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m8wzp/crc-debug-gv5q8" Nov 23 07:30:55 crc kubenswrapper[4559]: I1123 07:30:55.145240 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzrph\" (UniqueName: \"kubernetes.io/projected/ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3-kube-api-access-zzrph\") pod \"crc-debug-gv5q8\" (UID: \"ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3\") " pod="openshift-must-gather-m8wzp/crc-debug-gv5q8" Nov 23 07:30:55 crc kubenswrapper[4559]: I1123 07:30:55.145422 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3-host\") pod \"crc-debug-gv5q8\" (UID: \"ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3\") " pod="openshift-must-gather-m8wzp/crc-debug-gv5q8" Nov 23 07:30:55 crc kubenswrapper[4559]: I1123 07:30:55.246272 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3-host\") pod \"crc-debug-gv5q8\" (UID: \"ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3\") " pod="openshift-must-gather-m8wzp/crc-debug-gv5q8" Nov 23 07:30:55 crc kubenswrapper[4559]: I1123 07:30:55.246397 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3-host\") pod \"crc-debug-gv5q8\" (UID: \"ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3\") " pod="openshift-must-gather-m8wzp/crc-debug-gv5q8" Nov 23 07:30:55 crc kubenswrapper[4559]: I1123 07:30:55.246403 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzrph\" (UniqueName: \"kubernetes.io/projected/ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3-kube-api-access-zzrph\") pod \"crc-debug-gv5q8\" (UID: \"ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3\") " pod="openshift-must-gather-m8wzp/crc-debug-gv5q8" Nov 23 07:30:55 crc kubenswrapper[4559]: I1123 07:30:55.260940 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzrph\" (UniqueName: \"kubernetes.io/projected/ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3-kube-api-access-zzrph\") pod \"crc-debug-gv5q8\" (UID: \"ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3\") " pod="openshift-must-gather-m8wzp/crc-debug-gv5q8" Nov 23 07:30:55 crc kubenswrapper[4559]: I1123 07:30:55.357071 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m8wzp/crc-debug-gv5q8" Nov 23 07:30:55 crc kubenswrapper[4559]: W1123 07:30:55.762577 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podffdd0197_ace4_42a8_9ccc_7a37ff23c2c3.slice/crio-4bf4376fb9b52f3e50bc811998b93b95334d953a90248d0578b5acfe08f9fe34 WatchSource:0}: Error finding container 4bf4376fb9b52f3e50bc811998b93b95334d953a90248d0578b5acfe08f9fe34: Status 404 returned error can't find the container with id 4bf4376fb9b52f3e50bc811998b93b95334d953a90248d0578b5acfe08f9fe34 Nov 23 07:30:55 crc kubenswrapper[4559]: I1123 07:30:55.903713 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m8wzp/crc-debug-n625q" Nov 23 07:30:55 crc kubenswrapper[4559]: I1123 07:30:55.955818 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gn94\" (UniqueName: \"kubernetes.io/projected/be3a330c-2918-4c87-89ae-2565413a2f72-kube-api-access-5gn94\") pod \"be3a330c-2918-4c87-89ae-2565413a2f72\" (UID: \"be3a330c-2918-4c87-89ae-2565413a2f72\") " Nov 23 07:30:55 crc kubenswrapper[4559]: I1123 07:30:55.956055 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/be3a330c-2918-4c87-89ae-2565413a2f72-host\") pod \"be3a330c-2918-4c87-89ae-2565413a2f72\" (UID: \"be3a330c-2918-4c87-89ae-2565413a2f72\") " Nov 23 07:30:55 crc kubenswrapper[4559]: I1123 07:30:55.956097 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/be3a330c-2918-4c87-89ae-2565413a2f72-host" (OuterVolumeSpecName: "host") pod "be3a330c-2918-4c87-89ae-2565413a2f72" (UID: "be3a330c-2918-4c87-89ae-2565413a2f72"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:30:55 crc kubenswrapper[4559]: I1123 07:30:55.956594 4559 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/be3a330c-2918-4c87-89ae-2565413a2f72-host\") on node \"crc\" DevicePath \"\"" Nov 23 07:30:55 crc kubenswrapper[4559]: I1123 07:30:55.958914 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be3a330c-2918-4c87-89ae-2565413a2f72-kube-api-access-5gn94" (OuterVolumeSpecName: "kube-api-access-5gn94") pod "be3a330c-2918-4c87-89ae-2565413a2f72" (UID: "be3a330c-2918-4c87-89ae-2565413a2f72"). InnerVolumeSpecName "kube-api-access-5gn94". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:30:56 crc kubenswrapper[4559]: I1123 07:30:56.058242 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gn94\" (UniqueName: \"kubernetes.io/projected/be3a330c-2918-4c87-89ae-2565413a2f72-kube-api-access-5gn94\") on node \"crc\" DevicePath \"\"" Nov 23 07:30:56 crc kubenswrapper[4559]: I1123 07:30:56.281485 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be3a330c-2918-4c87-89ae-2565413a2f72" path="/var/lib/kubelet/pods/be3a330c-2918-4c87-89ae-2565413a2f72/volumes" Nov 23 07:30:56 crc kubenswrapper[4559]: I1123 07:30:56.465797 4559 scope.go:117] "RemoveContainer" containerID="553a9f1601de390f4e7885031dc7f1165e83459f56bac5616affd7f3aa19a499" Nov 23 07:30:56 crc kubenswrapper[4559]: I1123 07:30:56.465837 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m8wzp/crc-debug-n625q" Nov 23 07:30:56 crc kubenswrapper[4559]: I1123 07:30:56.467623 4559 generic.go:334] "Generic (PLEG): container finished" podID="c2daeb5e-241f-4764-962b-64c9c3d34597" containerID="d333466974f444cd27e777363ff02c6bd4564963d62a340fff552a8f9a687b6e" exitCode=0 Nov 23 07:30:56 crc kubenswrapper[4559]: I1123 07:30:56.467733 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nw54h" event={"ID":"c2daeb5e-241f-4764-962b-64c9c3d34597","Type":"ContainerDied","Data":"d333466974f444cd27e777363ff02c6bd4564963d62a340fff552a8f9a687b6e"} Nov 23 07:30:56 crc kubenswrapper[4559]: I1123 07:30:56.469551 4559 generic.go:334] "Generic (PLEG): container finished" podID="ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3" containerID="e774f3961f39e500d9e40565a4edbcf5d2c282a6ef02bc208cfe443640d5bd2e" exitCode=0 Nov 23 07:30:56 crc kubenswrapper[4559]: I1123 07:30:56.469574 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m8wzp/crc-debug-gv5q8" event={"ID":"ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3","Type":"ContainerDied","Data":"e774f3961f39e500d9e40565a4edbcf5d2c282a6ef02bc208cfe443640d5bd2e"} Nov 23 07:30:56 crc kubenswrapper[4559]: I1123 07:30:56.469595 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m8wzp/crc-debug-gv5q8" event={"ID":"ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3","Type":"ContainerStarted","Data":"4bf4376fb9b52f3e50bc811998b93b95334d953a90248d0578b5acfe08f9fe34"} Nov 23 07:30:56 crc kubenswrapper[4559]: I1123 07:30:56.509291 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-m8wzp/crc-debug-gv5q8"] Nov 23 07:30:56 crc kubenswrapper[4559]: I1123 07:30:56.516391 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-m8wzp/crc-debug-gv5q8"] Nov 23 07:30:57 crc kubenswrapper[4559]: I1123 07:30:57.477039 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nw54h" event={"ID":"c2daeb5e-241f-4764-962b-64c9c3d34597","Type":"ContainerStarted","Data":"c4d1b5c3d4c1833a460a2a101810b15f84354ce2817409880b73d4fd047b229f"} Nov 23 07:30:57 crc kubenswrapper[4559]: I1123 07:30:57.500380 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nw54h" podStartSLOduration=2.001987669 podStartE2EDuration="7.50036556s" podCreationTimestamp="2025-11-23 07:30:50 +0000 UTC" firstStartedPulling="2025-11-23 07:30:51.435914193 +0000 UTC m=+2753.457899807" lastFinishedPulling="2025-11-23 07:30:56.934292083 +0000 UTC m=+2758.956277698" observedRunningTime="2025-11-23 07:30:57.497220758 +0000 UTC m=+2759.519206372" watchObservedRunningTime="2025-11-23 07:30:57.50036556 +0000 UTC m=+2759.522351173" Nov 23 07:30:57 crc kubenswrapper[4559]: I1123 07:30:57.547716 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m8wzp/crc-debug-gv5q8" Nov 23 07:30:57 crc kubenswrapper[4559]: I1123 07:30:57.683427 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3-host\") pod \"ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3\" (UID: \"ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3\") " Nov 23 07:30:57 crc kubenswrapper[4559]: I1123 07:30:57.683508 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzrph\" (UniqueName: \"kubernetes.io/projected/ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3-kube-api-access-zzrph\") pod \"ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3\" (UID: \"ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3\") " Nov 23 07:30:57 crc kubenswrapper[4559]: I1123 07:30:57.683816 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3-host" (OuterVolumeSpecName: "host") pod "ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3" (UID: "ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:30:57 crc kubenswrapper[4559]: I1123 07:30:57.684097 4559 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3-host\") on node \"crc\" DevicePath \"\"" Nov 23 07:30:57 crc kubenswrapper[4559]: I1123 07:30:57.688086 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3-kube-api-access-zzrph" (OuterVolumeSpecName: "kube-api-access-zzrph") pod "ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3" (UID: "ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3"). InnerVolumeSpecName "kube-api-access-zzrph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:30:57 crc kubenswrapper[4559]: I1123 07:30:57.785542 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzrph\" (UniqueName: \"kubernetes.io/projected/ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3-kube-api-access-zzrph\") on node \"crc\" DevicePath \"\"" Nov 23 07:30:58 crc kubenswrapper[4559]: I1123 07:30:58.285258 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3" path="/var/lib/kubelet/pods/ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3/volumes" Nov 23 07:30:58 crc kubenswrapper[4559]: I1123 07:30:58.493451 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m8wzp/crc-debug-gv5q8" Nov 23 07:30:58 crc kubenswrapper[4559]: I1123 07:30:58.493487 4559 scope.go:117] "RemoveContainer" containerID="e774f3961f39e500d9e40565a4edbcf5d2c282a6ef02bc208cfe443640d5bd2e" Nov 23 07:31:00 crc kubenswrapper[4559]: I1123 07:31:00.470243 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nw54h" Nov 23 07:31:00 crc kubenswrapper[4559]: I1123 07:31:00.470464 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nw54h" Nov 23 07:31:00 crc kubenswrapper[4559]: I1123 07:31:00.503133 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nw54h" Nov 23 07:31:04 crc kubenswrapper[4559]: E1123 07:31:04.844677 4559 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podffdd0197_ace4_42a8_9ccc_7a37ff23c2c3.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podffdd0197_ace4_42a8_9ccc_7a37ff23c2c3.slice/crio-4bf4376fb9b52f3e50bc811998b93b95334d953a90248d0578b5acfe08f9fe34\": RecentStats: unable to find data in memory cache]" Nov 23 07:31:07 crc kubenswrapper[4559]: I1123 07:31:07.149961 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-55bb64d698-6lpkd_f9264f09-464d-404a-88a1-7f86ecb2ccb8/barbican-api/0.log" Nov 23 07:31:07 crc kubenswrapper[4559]: I1123 07:31:07.223849 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-55bb64d698-6lpkd_f9264f09-464d-404a-88a1-7f86ecb2ccb8/barbican-api-log/0.log" Nov 23 07:31:07 crc kubenswrapper[4559]: I1123 07:31:07.303300 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-685df6659d-zwp5c_cb7ba735-691f-4eb5-ac0c-c3f976e8e604/barbican-keystone-listener/0.log" Nov 23 07:31:07 crc kubenswrapper[4559]: I1123 07:31:07.335466 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-685df6659d-zwp5c_cb7ba735-691f-4eb5-ac0c-c3f976e8e604/barbican-keystone-listener-log/0.log" Nov 23 07:31:07 crc kubenswrapper[4559]: I1123 07:31:07.487531 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-d4dd9b9cc-ndjpd_f5b2ad32-159b-4369-951d-02d3ae8581eb/barbican-worker-log/0.log" Nov 23 07:31:07 crc kubenswrapper[4559]: I1123 07:31:07.498353 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-d4dd9b9cc-ndjpd_f5b2ad32-159b-4369-951d-02d3ae8581eb/barbican-worker/0.log" Nov 23 07:31:07 crc kubenswrapper[4559]: I1123 07:31:07.605548 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz_af59bdf5-b08e-487d-bcc4-c70a275cace5/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:31:07 crc kubenswrapper[4559]: I1123 07:31:07.675270 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_865e8bed-9dfe-4f61-9c40-e2876af8f95b/ceilometer-central-agent/0.log" Nov 23 07:31:07 crc kubenswrapper[4559]: I1123 07:31:07.744816 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_865e8bed-9dfe-4f61-9c40-e2876af8f95b/ceilometer-notification-agent/0.log" Nov 23 07:31:07 crc kubenswrapper[4559]: I1123 07:31:07.793432 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_865e8bed-9dfe-4f61-9c40-e2876af8f95b/proxy-httpd/0.log" Nov 23 07:31:07 crc kubenswrapper[4559]: I1123 07:31:07.814691 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_865e8bed-9dfe-4f61-9c40-e2876af8f95b/sg-core/0.log" Nov 23 07:31:07 crc kubenswrapper[4559]: I1123 07:31:07.916859 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_3b8ac924-b4b7-4374-8f5f-7b2133936385/cinder-api/0.log" Nov 23 07:31:07 crc kubenswrapper[4559]: I1123 07:31:07.938357 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_3b8ac924-b4b7-4374-8f5f-7b2133936385/cinder-api-log/0.log" Nov 23 07:31:08 crc kubenswrapper[4559]: I1123 07:31:08.054110 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_cd7f1662-fa90-4677-9c1a-e5882f079497/cinder-scheduler/0.log" Nov 23 07:31:08 crc kubenswrapper[4559]: I1123 07:31:08.084973 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_cd7f1662-fa90-4677-9c1a-e5882f079497/probe/0.log" Nov 23 07:31:08 crc kubenswrapper[4559]: I1123 07:31:08.185055 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5_45c516df-85dd-4a43-94ef-ea1381622c06/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:31:08 crc kubenswrapper[4559]: I1123 07:31:08.229518 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms_f3cbccfd-3b9d-4d8d-aec1-702b2e494667/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:31:08 crc kubenswrapper[4559]: I1123 07:31:08.337124 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-869db7cd47-ghbfq_9b236045-a74d-4a5e-a8e3-c70530d9e6a2/init/0.log" Nov 23 07:31:08 crc kubenswrapper[4559]: I1123 07:31:08.466829 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-869db7cd47-ghbfq_9b236045-a74d-4a5e-a8e3-c70530d9e6a2/init/0.log" Nov 23 07:31:08 crc kubenswrapper[4559]: I1123 07:31:08.495192 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-46r2s_fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:31:08 crc kubenswrapper[4559]: I1123 07:31:08.541211 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-869db7cd47-ghbfq_9b236045-a74d-4a5e-a8e3-c70530d9e6a2/dnsmasq-dns/0.log" Nov 23 07:31:08 crc kubenswrapper[4559]: I1123 07:31:08.670422 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_24caf856-63a1-44eb-a0f2-7afc985ff668/glance-httpd/0.log" Nov 23 07:31:08 crc kubenswrapper[4559]: I1123 07:31:08.707476 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_24caf856-63a1-44eb-a0f2-7afc985ff668/glance-log/0.log" Nov 23 07:31:08 crc kubenswrapper[4559]: I1123 07:31:08.792816 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_c869f924-be1d-4e33-8a16-5a833a6ad9d0/glance-httpd/0.log" Nov 23 07:31:08 crc kubenswrapper[4559]: I1123 07:31:08.821561 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_c869f924-be1d-4e33-8a16-5a833a6ad9d0/glance-log/0.log" Nov 23 07:31:08 crc kubenswrapper[4559]: I1123 07:31:08.896522 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs_aa380021-cb24-43b1-bb0d-e3dbbd5f6676/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:31:09 crc kubenswrapper[4559]: I1123 07:31:09.006769 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-wgnh6_460a9b32-addb-458a-9913-4ba0433b3eb3/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:31:09 crc kubenswrapper[4559]: I1123 07:31:09.118171 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-74ff86f86c-gcbcl_dca8804f-80bf-4b44-99d0-7c2a9fea9166/keystone-api/0.log" Nov 23 07:31:09 crc kubenswrapper[4559]: I1123 07:31:09.181278 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29398021-r88wd_da23948d-f4d4-42cc-9960-aaeb3efc2309/keystone-cron/0.log" Nov 23 07:31:09 crc kubenswrapper[4559]: I1123 07:31:09.267817 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_916e4424-5f09-44c7-8b07-de2a4d84df18/kube-state-metrics/0.log" Nov 23 07:31:09 crc kubenswrapper[4559]: I1123 07:31:09.435210 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-njqpj_aa33c87f-c563-4944-883e-3ed5649b96a5/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:31:09 crc kubenswrapper[4559]: I1123 07:31:09.679633 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-d8c98f585-kr7vt_72027df5-f460-436a-b883-1895caea6f90/neutron-httpd/0.log" Nov 23 07:31:09 crc kubenswrapper[4559]: I1123 07:31:09.720860 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-d8c98f585-kr7vt_72027df5-f460-436a-b883-1895caea6f90/neutron-api/0.log" Nov 23 07:31:09 crc kubenswrapper[4559]: I1123 07:31:09.944543 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n_4f9b53ba-da26-40bb-9819-cdeb54deaef2/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:31:10 crc kubenswrapper[4559]: I1123 07:31:10.286856 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_a5666b61-0dcd-4fcd-a96a-44fa29cd3d84/nova-cell0-conductor-conductor/0.log" Nov 23 07:31:10 crc kubenswrapper[4559]: I1123 07:31:10.306602 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e55ca2b9-735d-434f-899d-cef9bce42b2e/nova-api-log/0.log" Nov 23 07:31:10 crc kubenswrapper[4559]: I1123 07:31:10.413040 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e55ca2b9-735d-434f-899d-cef9bce42b2e/nova-api-api/0.log" Nov 23 07:31:10 crc kubenswrapper[4559]: I1123 07:31:10.509300 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nw54h" Nov 23 07:31:10 crc kubenswrapper[4559]: I1123 07:31:10.569089 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b/nova-cell1-novncproxy-novncproxy/0.log" Nov 23 07:31:10 crc kubenswrapper[4559]: I1123 07:31:10.574655 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nw54h"] Nov 23 07:31:10 crc kubenswrapper[4559]: I1123 07:31:10.600516 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_56c491cd-4182-4477-a9de-d7e91d68e07f/nova-cell1-conductor-conductor/0.log" Nov 23 07:31:10 crc kubenswrapper[4559]: I1123 07:31:10.602702 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6mnmc"] Nov 23 07:31:10 crc kubenswrapper[4559]: I1123 07:31:10.602999 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6mnmc" podUID="23ec849f-d94a-45b0-9f74-655f33ffb163" containerName="registry-server" containerID="cri-o://6b00c16d84a1cad97b06f59023b6089017c02695da06a30ef74711818044c86e" gracePeriod=2 Nov 23 07:31:10 crc kubenswrapper[4559]: I1123 07:31:10.833518 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-dfw5h_2505b564-8639-464b-b223-91af7ab3661d/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:31:10 crc kubenswrapper[4559]: I1123 07:31:10.994029 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_66d6e144-f771-40b4-a6f2-0689654cf797/nova-metadata-log/0.log" Nov 23 07:31:11 crc kubenswrapper[4559]: I1123 07:31:11.067072 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6mnmc" Nov 23 07:31:11 crc kubenswrapper[4559]: I1123 07:31:11.107705 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpgz5\" (UniqueName: \"kubernetes.io/projected/23ec849f-d94a-45b0-9f74-655f33ffb163-kube-api-access-hpgz5\") pod \"23ec849f-d94a-45b0-9f74-655f33ffb163\" (UID: \"23ec849f-d94a-45b0-9f74-655f33ffb163\") " Nov 23 07:31:11 crc kubenswrapper[4559]: I1123 07:31:11.107788 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23ec849f-d94a-45b0-9f74-655f33ffb163-utilities\") pod \"23ec849f-d94a-45b0-9f74-655f33ffb163\" (UID: \"23ec849f-d94a-45b0-9f74-655f33ffb163\") " Nov 23 07:31:11 crc kubenswrapper[4559]: I1123 07:31:11.107826 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23ec849f-d94a-45b0-9f74-655f33ffb163-catalog-content\") pod \"23ec849f-d94a-45b0-9f74-655f33ffb163\" (UID: \"23ec849f-d94a-45b0-9f74-655f33ffb163\") " Nov 23 07:31:11 crc kubenswrapper[4559]: I1123 07:31:11.110010 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23ec849f-d94a-45b0-9f74-655f33ffb163-utilities" (OuterVolumeSpecName: "utilities") pod "23ec849f-d94a-45b0-9f74-655f33ffb163" (UID: "23ec849f-d94a-45b0-9f74-655f33ffb163"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:31:11 crc kubenswrapper[4559]: I1123 07:31:11.119760 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23ec849f-d94a-45b0-9f74-655f33ffb163-kube-api-access-hpgz5" (OuterVolumeSpecName: "kube-api-access-hpgz5") pod "23ec849f-d94a-45b0-9f74-655f33ffb163" (UID: "23ec849f-d94a-45b0-9f74-655f33ffb163"). InnerVolumeSpecName "kube-api-access-hpgz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:31:11 crc kubenswrapper[4559]: I1123 07:31:11.209066 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23ec849f-d94a-45b0-9f74-655f33ffb163-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "23ec849f-d94a-45b0-9f74-655f33ffb163" (UID: "23ec849f-d94a-45b0-9f74-655f33ffb163"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:31:11 crc kubenswrapper[4559]: I1123 07:31:11.211490 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23ec849f-d94a-45b0-9f74-655f33ffb163-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:31:11 crc kubenswrapper[4559]: I1123 07:31:11.211520 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpgz5\" (UniqueName: \"kubernetes.io/projected/23ec849f-d94a-45b0-9f74-655f33ffb163-kube-api-access-hpgz5\") on node \"crc\" DevicePath \"\"" Nov 23 07:31:11 crc kubenswrapper[4559]: I1123 07:31:11.211535 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23ec849f-d94a-45b0-9f74-655f33ffb163-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:31:11 crc kubenswrapper[4559]: I1123 07:31:11.601893 4559 generic.go:334] "Generic (PLEG): container finished" podID="23ec849f-d94a-45b0-9f74-655f33ffb163" containerID="6b00c16d84a1cad97b06f59023b6089017c02695da06a30ef74711818044c86e" exitCode=0 Nov 23 07:31:11 crc kubenswrapper[4559]: I1123 07:31:11.601952 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6mnmc" event={"ID":"23ec849f-d94a-45b0-9f74-655f33ffb163","Type":"ContainerDied","Data":"6b00c16d84a1cad97b06f59023b6089017c02695da06a30ef74711818044c86e"} Nov 23 07:31:11 crc kubenswrapper[4559]: I1123 07:31:11.602743 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6mnmc" Nov 23 07:31:11 crc kubenswrapper[4559]: I1123 07:31:11.601976 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6mnmc" event={"ID":"23ec849f-d94a-45b0-9f74-655f33ffb163","Type":"ContainerDied","Data":"16b68db09c6d1d2bbc74e64a3e61102d7bd957ae3993365edf0a977025e347dc"} Nov 23 07:31:11 crc kubenswrapper[4559]: I1123 07:31:11.602906 4559 scope.go:117] "RemoveContainer" containerID="6b00c16d84a1cad97b06f59023b6089017c02695da06a30ef74711818044c86e" Nov 23 07:31:11 crc kubenswrapper[4559]: I1123 07:31:11.636118 4559 scope.go:117] "RemoveContainer" containerID="1973e1efebe3c33a41615a39f1f0a95af29d8087ad7606cfac7e95d9f0fda64f" Nov 23 07:31:11 crc kubenswrapper[4559]: I1123 07:31:11.640739 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6mnmc"] Nov 23 07:31:11 crc kubenswrapper[4559]: I1123 07:31:11.645758 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_01dff7d1-bf4c-4d88-900c-99de62949ac7/mysql-bootstrap/0.log" Nov 23 07:31:11 crc kubenswrapper[4559]: I1123 07:31:11.660006 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6mnmc"] Nov 23 07:31:11 crc kubenswrapper[4559]: I1123 07:31:11.666415 4559 scope.go:117] "RemoveContainer" containerID="d51204fb5c5393526d79c270481975d9d23198496963a689b1ed0ffbb3763c0e" Nov 23 07:31:11 crc kubenswrapper[4559]: I1123 07:31:11.687683 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_ba80f653-69d6-4a80-aaab-2747e0f9b61b/nova-scheduler-scheduler/0.log" Nov 23 07:31:11 crc kubenswrapper[4559]: I1123 07:31:11.703831 4559 scope.go:117] "RemoveContainer" containerID="6b00c16d84a1cad97b06f59023b6089017c02695da06a30ef74711818044c86e" Nov 23 07:31:11 crc kubenswrapper[4559]: E1123 07:31:11.704236 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b00c16d84a1cad97b06f59023b6089017c02695da06a30ef74711818044c86e\": container with ID starting with 6b00c16d84a1cad97b06f59023b6089017c02695da06a30ef74711818044c86e not found: ID does not exist" containerID="6b00c16d84a1cad97b06f59023b6089017c02695da06a30ef74711818044c86e" Nov 23 07:31:11 crc kubenswrapper[4559]: I1123 07:31:11.704322 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b00c16d84a1cad97b06f59023b6089017c02695da06a30ef74711818044c86e"} err="failed to get container status \"6b00c16d84a1cad97b06f59023b6089017c02695da06a30ef74711818044c86e\": rpc error: code = NotFound desc = could not find container \"6b00c16d84a1cad97b06f59023b6089017c02695da06a30ef74711818044c86e\": container with ID starting with 6b00c16d84a1cad97b06f59023b6089017c02695da06a30ef74711818044c86e not found: ID does not exist" Nov 23 07:31:11 crc kubenswrapper[4559]: I1123 07:31:11.704369 4559 scope.go:117] "RemoveContainer" containerID="1973e1efebe3c33a41615a39f1f0a95af29d8087ad7606cfac7e95d9f0fda64f" Nov 23 07:31:11 crc kubenswrapper[4559]: E1123 07:31:11.704690 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1973e1efebe3c33a41615a39f1f0a95af29d8087ad7606cfac7e95d9f0fda64f\": container with ID starting with 1973e1efebe3c33a41615a39f1f0a95af29d8087ad7606cfac7e95d9f0fda64f not found: ID does not exist" containerID="1973e1efebe3c33a41615a39f1f0a95af29d8087ad7606cfac7e95d9f0fda64f" Nov 23 07:31:11 crc kubenswrapper[4559]: I1123 07:31:11.704735 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1973e1efebe3c33a41615a39f1f0a95af29d8087ad7606cfac7e95d9f0fda64f"} err="failed to get container status \"1973e1efebe3c33a41615a39f1f0a95af29d8087ad7606cfac7e95d9f0fda64f\": rpc error: code = NotFound desc = could not find container \"1973e1efebe3c33a41615a39f1f0a95af29d8087ad7606cfac7e95d9f0fda64f\": container with ID starting with 1973e1efebe3c33a41615a39f1f0a95af29d8087ad7606cfac7e95d9f0fda64f not found: ID does not exist" Nov 23 07:31:11 crc kubenswrapper[4559]: I1123 07:31:11.704766 4559 scope.go:117] "RemoveContainer" containerID="d51204fb5c5393526d79c270481975d9d23198496963a689b1ed0ffbb3763c0e" Nov 23 07:31:11 crc kubenswrapper[4559]: E1123 07:31:11.709802 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d51204fb5c5393526d79c270481975d9d23198496963a689b1ed0ffbb3763c0e\": container with ID starting with d51204fb5c5393526d79c270481975d9d23198496963a689b1ed0ffbb3763c0e not found: ID does not exist" containerID="d51204fb5c5393526d79c270481975d9d23198496963a689b1ed0ffbb3763c0e" Nov 23 07:31:11 crc kubenswrapper[4559]: I1123 07:31:11.709837 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d51204fb5c5393526d79c270481975d9d23198496963a689b1ed0ffbb3763c0e"} err="failed to get container status \"d51204fb5c5393526d79c270481975d9d23198496963a689b1ed0ffbb3763c0e\": rpc error: code = NotFound desc = could not find container \"d51204fb5c5393526d79c270481975d9d23198496963a689b1ed0ffbb3763c0e\": container with ID starting with d51204fb5c5393526d79c270481975d9d23198496963a689b1ed0ffbb3763c0e not found: ID does not exist" Nov 23 07:31:11 crc kubenswrapper[4559]: I1123 07:31:11.803459 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_66d6e144-f771-40b4-a6f2-0689654cf797/nova-metadata-metadata/0.log" Nov 23 07:31:12 crc kubenswrapper[4559]: I1123 07:31:12.068531 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_01dff7d1-bf4c-4d88-900c-99de62949ac7/galera/0.log" Nov 23 07:31:12 crc kubenswrapper[4559]: I1123 07:31:12.099283 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_01dff7d1-bf4c-4d88-900c-99de62949ac7/mysql-bootstrap/0.log" Nov 23 07:31:12 crc kubenswrapper[4559]: I1123 07:31:12.103726 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e097ab60-4653-4038-b6de-40b55936565c/mysql-bootstrap/0.log" Nov 23 07:31:12 crc kubenswrapper[4559]: I1123 07:31:12.283508 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23ec849f-d94a-45b0-9f74-655f33ffb163" path="/var/lib/kubelet/pods/23ec849f-d94a-45b0-9f74-655f33ffb163/volumes" Nov 23 07:31:12 crc kubenswrapper[4559]: I1123 07:31:12.463265 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e097ab60-4653-4038-b6de-40b55936565c/mysql-bootstrap/0.log" Nov 23 07:31:12 crc kubenswrapper[4559]: I1123 07:31:12.550327 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e097ab60-4653-4038-b6de-40b55936565c/galera/0.log" Nov 23 07:31:12 crc kubenswrapper[4559]: I1123 07:31:12.642289 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa/openstackclient/0.log" Nov 23 07:31:12 crc kubenswrapper[4559]: I1123 07:31:12.724450 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-blj85_20662e83-ba79-4c97-80fb-98fda28c1149/ovn-controller/0.log" Nov 23 07:31:12 crc kubenswrapper[4559]: I1123 07:31:12.807332 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-nzw6k_dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5/openstack-network-exporter/0.log" Nov 23 07:31:12 crc kubenswrapper[4559]: I1123 07:31:12.928621 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jg7hk_d534f227-972b-4a85-aeb5-8f1d226352e1/ovsdb-server-init/0.log" Nov 23 07:31:13 crc kubenswrapper[4559]: I1123 07:31:13.091509 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jg7hk_d534f227-972b-4a85-aeb5-8f1d226352e1/ovs-vswitchd/0.log" Nov 23 07:31:13 crc kubenswrapper[4559]: I1123 07:31:13.099957 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jg7hk_d534f227-972b-4a85-aeb5-8f1d226352e1/ovsdb-server-init/0.log" Nov 23 07:31:13 crc kubenswrapper[4559]: I1123 07:31:13.110796 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jg7hk_d534f227-972b-4a85-aeb5-8f1d226352e1/ovsdb-server/0.log" Nov 23 07:31:13 crc kubenswrapper[4559]: I1123 07:31:13.287175 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-9svlh_4f4ac16e-c282-4fa3-bdbe-e5f848782f47/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:31:13 crc kubenswrapper[4559]: I1123 07:31:13.313009 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_00fa46d4-5d8e-43a6-a182-34faf0d694ab/openstack-network-exporter/0.log" Nov 23 07:31:13 crc kubenswrapper[4559]: I1123 07:31:13.339076 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_00fa46d4-5d8e-43a6-a182-34faf0d694ab/ovn-northd/0.log" Nov 23 07:31:13 crc kubenswrapper[4559]: I1123 07:31:13.520337 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_5d19fe8b-239a-476a-9213-1eccbd58958e/openstack-network-exporter/0.log" Nov 23 07:31:13 crc kubenswrapper[4559]: I1123 07:31:13.564617 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_5d19fe8b-239a-476a-9213-1eccbd58958e/ovsdbserver-nb/0.log" Nov 23 07:31:13 crc kubenswrapper[4559]: I1123 07:31:13.681447 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f162538d-e942-4d5e-b2c5-11ba57054cf4/ovsdbserver-sb/0.log" Nov 23 07:31:13 crc kubenswrapper[4559]: I1123 07:31:13.690345 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f162538d-e942-4d5e-b2c5-11ba57054cf4/openstack-network-exporter/0.log" Nov 23 07:31:13 crc kubenswrapper[4559]: I1123 07:31:13.826063 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7667f44776-mszxt_1df68d96-dfae-42c9-8a11-9906486ea1b9/placement-api/0.log" Nov 23 07:31:13 crc kubenswrapper[4559]: I1123 07:31:13.934463 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_e0731be9-7300-45df-b028-8aeb2a3999b4/setup-container/0.log" Nov 23 07:31:13 crc kubenswrapper[4559]: I1123 07:31:13.991674 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7667f44776-mszxt_1df68d96-dfae-42c9-8a11-9906486ea1b9/placement-log/0.log" Nov 23 07:31:14 crc kubenswrapper[4559]: I1123 07:31:14.111862 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_e0731be9-7300-45df-b028-8aeb2a3999b4/setup-container/0.log" Nov 23 07:31:14 crc kubenswrapper[4559]: I1123 07:31:14.157787 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8/setup-container/0.log" Nov 23 07:31:14 crc kubenswrapper[4559]: I1123 07:31:14.158381 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_e0731be9-7300-45df-b028-8aeb2a3999b4/rabbitmq/0.log" Nov 23 07:31:14 crc kubenswrapper[4559]: I1123 07:31:14.387105 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8/setup-container/0.log" Nov 23 07:31:14 crc kubenswrapper[4559]: I1123 07:31:14.472815 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8/rabbitmq/0.log" Nov 23 07:31:14 crc kubenswrapper[4559]: I1123 07:31:14.522358 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2_67ea0e44-543c-4601-88e5-51ee0d43424a/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:31:14 crc kubenswrapper[4559]: I1123 07:31:14.614573 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-sx224_1b84c748-5838-4938-be55-0e8fc4174730/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:31:14 crc kubenswrapper[4559]: I1123 07:31:14.727544 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb_c5fa1816-8789-4600-931c-8a376f0936c0/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:31:14 crc kubenswrapper[4559]: I1123 07:31:14.854108 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-5flwd_f91e51c1-c500-4da2-acba-fcdfa134d397/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:31:14 crc kubenswrapper[4559]: I1123 07:31:14.900500 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-sw6gw_3f1151cd-5698-4fd5-ba25-e28b48717cb8/ssh-known-hosts-edpm-deployment/0.log" Nov 23 07:31:15 crc kubenswrapper[4559]: E1123 07:31:15.053249 4559 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podffdd0197_ace4_42a8_9ccc_7a37ff23c2c3.slice/crio-4bf4376fb9b52f3e50bc811998b93b95334d953a90248d0578b5acfe08f9fe34\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podffdd0197_ace4_42a8_9ccc_7a37ff23c2c3.slice\": RecentStats: unable to find data in memory cache]" Nov 23 07:31:15 crc kubenswrapper[4559]: I1123 07:31:15.118380 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6b4fdd88f9-9hnng_db083799-cb94-42d3-8a54-8446f7a76502/proxy-server/0.log" Nov 23 07:31:15 crc kubenswrapper[4559]: I1123 07:31:15.134152 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6b4fdd88f9-9hnng_db083799-cb94-42d3-8a54-8446f7a76502/proxy-httpd/0.log" Nov 23 07:31:15 crc kubenswrapper[4559]: I1123 07:31:15.214256 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-94gfd_3696cad8-c9be-4efd-982a-a4c1d6de858b/swift-ring-rebalance/0.log" Nov 23 07:31:15 crc kubenswrapper[4559]: I1123 07:31:15.305623 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/account-auditor/0.log" Nov 23 07:31:15 crc kubenswrapper[4559]: I1123 07:31:15.369844 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/account-reaper/0.log" Nov 23 07:31:15 crc kubenswrapper[4559]: I1123 07:31:15.415949 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/account-replicator/0.log" Nov 23 07:31:15 crc kubenswrapper[4559]: I1123 07:31:15.505751 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/account-server/0.log" Nov 23 07:31:15 crc kubenswrapper[4559]: I1123 07:31:15.564719 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/container-auditor/0.log" Nov 23 07:31:15 crc kubenswrapper[4559]: I1123 07:31:15.596898 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/container-replicator/0.log" Nov 23 07:31:15 crc kubenswrapper[4559]: I1123 07:31:15.690736 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/container-server/0.log" Nov 23 07:31:15 crc kubenswrapper[4559]: I1123 07:31:15.748863 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/object-auditor/0.log" Nov 23 07:31:15 crc kubenswrapper[4559]: I1123 07:31:15.762344 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/container-updater/0.log" Nov 23 07:31:15 crc kubenswrapper[4559]: I1123 07:31:15.830337 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/object-expirer/0.log" Nov 23 07:31:15 crc kubenswrapper[4559]: I1123 07:31:15.882514 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/object-replicator/0.log" Nov 23 07:31:15 crc kubenswrapper[4559]: I1123 07:31:15.959297 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/object-server/0.log" Nov 23 07:31:15 crc kubenswrapper[4559]: I1123 07:31:15.960121 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/object-updater/0.log" Nov 23 07:31:16 crc kubenswrapper[4559]: I1123 07:31:16.077767 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/rsync/0.log" Nov 23 07:31:16 crc kubenswrapper[4559]: I1123 07:31:16.096329 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/swift-recon-cron/0.log" Nov 23 07:31:16 crc kubenswrapper[4559]: I1123 07:31:16.177980 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8_14f31bb2-9f00-4931-a84a-900401189cc0/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:31:16 crc kubenswrapper[4559]: I1123 07:31:16.353792 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_fb299f6e-9a30-4d29-95db-a0cbceb573ea/tempest-tests-tempest-tests-runner/0.log" Nov 23 07:31:16 crc kubenswrapper[4559]: I1123 07:31:16.361302 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_83a546c2-d73a-40cd-aae5-22eb4aef5268/test-operator-logs-container/0.log" Nov 23 07:31:16 crc kubenswrapper[4559]: I1123 07:31:16.536083 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-crmzz_2368491a-9b45-4c74-bb8d-b6be5fca89f8/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:31:22 crc kubenswrapper[4559]: I1123 07:31:22.953720 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_de0cb0c4-ffde-44c5-adb9-7aea0692c1b5/memcached/0.log" Nov 23 07:31:25 crc kubenswrapper[4559]: E1123 07:31:25.267037 4559 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podffdd0197_ace4_42a8_9ccc_7a37ff23c2c3.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podffdd0197_ace4_42a8_9ccc_7a37ff23c2c3.slice/crio-4bf4376fb9b52f3e50bc811998b93b95334d953a90248d0578b5acfe08f9fe34\": RecentStats: unable to find data in memory cache]" Nov 23 07:31:26 crc kubenswrapper[4559]: I1123 07:31:26.166489 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:31:26 crc kubenswrapper[4559]: I1123 07:31:26.166539 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:31:32 crc kubenswrapper[4559]: I1123 07:31:32.900465 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt_aaa08108-3039-4f4c-b01f-93dc7deefaa5/util/0.log" Nov 23 07:31:33 crc kubenswrapper[4559]: I1123 07:31:33.026327 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt_aaa08108-3039-4f4c-b01f-93dc7deefaa5/util/0.log" Nov 23 07:31:33 crc kubenswrapper[4559]: I1123 07:31:33.041367 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt_aaa08108-3039-4f4c-b01f-93dc7deefaa5/pull/0.log" Nov 23 07:31:33 crc kubenswrapper[4559]: I1123 07:31:33.059454 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt_aaa08108-3039-4f4c-b01f-93dc7deefaa5/pull/0.log" Nov 23 07:31:33 crc kubenswrapper[4559]: I1123 07:31:33.164424 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt_aaa08108-3039-4f4c-b01f-93dc7deefaa5/util/0.log" Nov 23 07:31:33 crc kubenswrapper[4559]: I1123 07:31:33.170099 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt_aaa08108-3039-4f4c-b01f-93dc7deefaa5/pull/0.log" Nov 23 07:31:33 crc kubenswrapper[4559]: I1123 07:31:33.184202 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt_aaa08108-3039-4f4c-b01f-93dc7deefaa5/extract/0.log" Nov 23 07:31:33 crc kubenswrapper[4559]: I1123 07:31:33.285345 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7768f8c84f-ks2b7_090f1d86-b4e0-43ca-9aee-49c771cab8c1/kube-rbac-proxy/0.log" Nov 23 07:31:33 crc kubenswrapper[4559]: I1123 07:31:33.347039 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7768f8c84f-ks2b7_090f1d86-b4e0-43ca-9aee-49c771cab8c1/manager/0.log" Nov 23 07:31:33 crc kubenswrapper[4559]: I1123 07:31:33.370573 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6d8fd67bf7-d78br_35944557-7948-4d54-8225-788d51eb01d1/kube-rbac-proxy/0.log" Nov 23 07:31:33 crc kubenswrapper[4559]: I1123 07:31:33.454722 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6d8fd67bf7-d78br_35944557-7948-4d54-8225-788d51eb01d1/manager/0.log" Nov 23 07:31:33 crc kubenswrapper[4559]: I1123 07:31:33.503842 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-56dfb6b67f-47bnx_6f79eaa5-12ee-4c47-8c5f-c0fa9190bc2d/kube-rbac-proxy/0.log" Nov 23 07:31:33 crc kubenswrapper[4559]: I1123 07:31:33.512040 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-56dfb6b67f-47bnx_6f79eaa5-12ee-4c47-8c5f-c0fa9190bc2d/manager/0.log" Nov 23 07:31:33 crc kubenswrapper[4559]: I1123 07:31:33.607347 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8667fbf6f6-qfgkh_042a59c3-a00e-4934-b16e-c731f0ab903f/kube-rbac-proxy/0.log" Nov 23 07:31:33 crc kubenswrapper[4559]: I1123 07:31:33.695699 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8667fbf6f6-qfgkh_042a59c3-a00e-4934-b16e-c731f0ab903f/manager/0.log" Nov 23 07:31:33 crc kubenswrapper[4559]: I1123 07:31:33.726403 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-bf4c6585d-9tvg2_5b70e018-1ab3-4f23-b43f-1f433a096b81/kube-rbac-proxy/0.log" Nov 23 07:31:33 crc kubenswrapper[4559]: I1123 07:31:33.753539 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-bf4c6585d-9tvg2_5b70e018-1ab3-4f23-b43f-1f433a096b81/manager/0.log" Nov 23 07:31:33 crc kubenswrapper[4559]: I1123 07:31:33.820282 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d86b44686-c776j_4cc947bb-881d-4592-9db6-ea53f6694ea5/kube-rbac-proxy/0.log" Nov 23 07:31:33 crc kubenswrapper[4559]: I1123 07:31:33.881446 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d86b44686-c776j_4cc947bb-881d-4592-9db6-ea53f6694ea5/manager/0.log" Nov 23 07:31:33 crc kubenswrapper[4559]: I1123 07:31:33.932017 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-769d9c7585-qq8sm_82279b3c-bca1-4891-b7d9-a367005ad84e/kube-rbac-proxy/0.log" Nov 23 07:31:34 crc kubenswrapper[4559]: I1123 07:31:34.006227 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5c75d7c94b-5tcjp_527a2a9e-741e-4f1b-8546-a852d73a836b/kube-rbac-proxy/0.log" Nov 23 07:31:34 crc kubenswrapper[4559]: I1123 07:31:34.068099 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-769d9c7585-qq8sm_82279b3c-bca1-4891-b7d9-a367005ad84e/manager/0.log" Nov 23 07:31:34 crc kubenswrapper[4559]: I1123 07:31:34.111551 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5c75d7c94b-5tcjp_527a2a9e-741e-4f1b-8546-a852d73a836b/manager/0.log" Nov 23 07:31:34 crc kubenswrapper[4559]: I1123 07:31:34.162994 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7879fb76fd-rxpxw_6935f82b-929a-421a-95d0-3315c9d0ef70/kube-rbac-proxy/0.log" Nov 23 07:31:34 crc kubenswrapper[4559]: I1123 07:31:34.229029 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7879fb76fd-rxpxw_6935f82b-929a-421a-95d0-3315c9d0ef70/manager/0.log" Nov 23 07:31:34 crc kubenswrapper[4559]: I1123 07:31:34.321970 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7bb88cb858-qgvkm_50ff7c3e-8c4a-45e9-abe0-f5391a5c7c95/kube-rbac-proxy/0.log" Nov 23 07:31:34 crc kubenswrapper[4559]: I1123 07:31:34.340041 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7bb88cb858-qgvkm_50ff7c3e-8c4a-45e9-abe0-f5391a5c7c95/manager/0.log" Nov 23 07:31:34 crc kubenswrapper[4559]: I1123 07:31:34.392309 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6f8c5b86cb-5rhv2_620fe37e-07bb-4286-b1e8-fd62db1cf022/kube-rbac-proxy/0.log" Nov 23 07:31:34 crc kubenswrapper[4559]: I1123 07:31:34.461092 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6f8c5b86cb-5rhv2_620fe37e-07bb-4286-b1e8-fd62db1cf022/manager/0.log" Nov 23 07:31:34 crc kubenswrapper[4559]: I1123 07:31:34.515919 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-66b7d6f598-hgbh8_7c53dd20-9115-4fac-a49e-b6b099150245/kube-rbac-proxy/0.log" Nov 23 07:31:34 crc kubenswrapper[4559]: I1123 07:31:34.544325 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-66b7d6f598-hgbh8_7c53dd20-9115-4fac-a49e-b6b099150245/manager/0.log" Nov 23 07:31:34 crc kubenswrapper[4559]: I1123 07:31:34.627210 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-86d796d84d-6ht8s_38f7bf7e-6d60-4cda-a988-ef9ce22883f8/kube-rbac-proxy/0.log" Nov 23 07:31:34 crc kubenswrapper[4559]: I1123 07:31:34.712187 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-86d796d84d-6ht8s_38f7bf7e-6d60-4cda-a988-ef9ce22883f8/manager/0.log" Nov 23 07:31:34 crc kubenswrapper[4559]: I1123 07:31:34.732913 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6fdc856c5d-bg99z_a659f54b-26d9-4681-963a-40348401f023/kube-rbac-proxy/0.log" Nov 23 07:31:34 crc kubenswrapper[4559]: I1123 07:31:34.778219 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6fdc856c5d-bg99z_a659f54b-26d9-4681-963a-40348401f023/manager/0.log" Nov 23 07:31:34 crc kubenswrapper[4559]: I1123 07:31:34.875226 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r_14445d59-81da-4b26-bd5d-2bbe2bd1a9c0/kube-rbac-proxy/0.log" Nov 23 07:31:34 crc kubenswrapper[4559]: I1123 07:31:34.884159 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r_14445d59-81da-4b26-bd5d-2bbe2bd1a9c0/manager/0.log" Nov 23 07:31:34 crc kubenswrapper[4559]: I1123 07:31:34.988135 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6cb9dc54f8-gmv99_ec988465-9f4e-46fe-beb3-e9032bf589d0/kube-rbac-proxy/0.log" Nov 23 07:31:35 crc kubenswrapper[4559]: I1123 07:31:35.096652 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-8486c7f98b-tfq8j_febcfb0d-e692-4e93-bbb5-40efe2929587/kube-rbac-proxy/0.log" Nov 23 07:31:35 crc kubenswrapper[4559]: I1123 07:31:35.325484 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-8486c7f98b-tfq8j_febcfb0d-e692-4e93-bbb5-40efe2929587/operator/0.log" Nov 23 07:31:35 crc kubenswrapper[4559]: I1123 07:31:35.394266 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-clkrn_7d86f9e5-a7b8-4807-bdeb-149b868a3c7a/registry-server/0.log" Nov 23 07:31:35 crc kubenswrapper[4559]: E1123 07:31:35.522671 4559 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podffdd0197_ace4_42a8_9ccc_7a37ff23c2c3.slice/crio-4bf4376fb9b52f3e50bc811998b93b95334d953a90248d0578b5acfe08f9fe34\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podffdd0197_ace4_42a8_9ccc_7a37ff23c2c3.slice\": RecentStats: unable to find data in memory cache]" Nov 23 07:31:35 crc kubenswrapper[4559]: I1123 07:31:35.719167 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5bdf4f7f7f-8hrnp_9c496b90-00da-4367-90f5-347a0c5a8ac6/kube-rbac-proxy/0.log" Nov 23 07:31:35 crc kubenswrapper[4559]: I1123 07:31:35.815798 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5bdf4f7f7f-8hrnp_9c496b90-00da-4367-90f5-347a0c5a8ac6/manager/0.log" Nov 23 07:31:35 crc kubenswrapper[4559]: I1123 07:31:35.831791 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6cb9dc54f8-gmv99_ec988465-9f4e-46fe-beb3-e9032bf589d0/manager/0.log" Nov 23 07:31:35 crc kubenswrapper[4559]: I1123 07:31:35.908376 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-6dc664666c-htkt7_c5127c1b-7342-4669-b04b-8484fd2ac326/kube-rbac-proxy/0.log" Nov 23 07:31:35 crc kubenswrapper[4559]: I1123 07:31:35.982465 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-6dc664666c-htkt7_c5127c1b-7342-4669-b04b-8484fd2ac326/manager/0.log" Nov 23 07:31:36 crc kubenswrapper[4559]: I1123 07:31:36.029326 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-f6s9r_465518b5-58ac-4347-aae0-94b8f68a6b36/operator/0.log" Nov 23 07:31:36 crc kubenswrapper[4559]: I1123 07:31:36.173976 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-799cb6ffd6-56qvk_b5998049-f03f-4de4-9601-dd970e3215fe/kube-rbac-proxy/0.log" Nov 23 07:31:36 crc kubenswrapper[4559]: I1123 07:31:36.186858 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-799cb6ffd6-56qvk_b5998049-f03f-4de4-9601-dd970e3215fe/manager/0.log" Nov 23 07:31:36 crc kubenswrapper[4559]: I1123 07:31:36.330688 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7798859c74-62tqn_6486d960-766f-4622-bad1-853e2b6a3fce/kube-rbac-proxy/0.log" Nov 23 07:31:36 crc kubenswrapper[4559]: I1123 07:31:36.335808 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7798859c74-62tqn_6486d960-766f-4622-bad1-853e2b6a3fce/manager/0.log" Nov 23 07:31:36 crc kubenswrapper[4559]: I1123 07:31:36.363238 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-8464cf66df-c2gjt_2463fe89-bc43-469f-9837-3ffd4b75605c/kube-rbac-proxy/0.log" Nov 23 07:31:36 crc kubenswrapper[4559]: I1123 07:31:36.399554 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-8464cf66df-c2gjt_2463fe89-bc43-469f-9837-3ffd4b75605c/manager/0.log" Nov 23 07:31:36 crc kubenswrapper[4559]: I1123 07:31:36.484778 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7cd4fb6f79-7brkl_a0f57f27-60d8-4fb8-ba76-c4ea5f3b99f6/kube-rbac-proxy/0.log" Nov 23 07:31:36 crc kubenswrapper[4559]: I1123 07:31:36.534613 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7cd4fb6f79-7brkl_a0f57f27-60d8-4fb8-ba76-c4ea5f3b99f6/manager/0.log" Nov 23 07:31:45 crc kubenswrapper[4559]: E1123 07:31:45.713591 4559 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podffdd0197_ace4_42a8_9ccc_7a37ff23c2c3.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podffdd0197_ace4_42a8_9ccc_7a37ff23c2c3.slice/crio-4bf4376fb9b52f3e50bc811998b93b95334d953a90248d0578b5acfe08f9fe34\": RecentStats: unable to find data in memory cache]" Nov 23 07:31:47 crc kubenswrapper[4559]: I1123 07:31:47.802020 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-2j2z7_ae746e9c-3187-4a3b-a439-b5ef25b37caf/control-plane-machine-set-operator/0.log" Nov 23 07:31:47 crc kubenswrapper[4559]: I1123 07:31:47.919975 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-ztnsp_18f99d77-1c2d-412b-93ae-1d1fc52f24ab/kube-rbac-proxy/0.log" Nov 23 07:31:47 crc kubenswrapper[4559]: I1123 07:31:47.948367 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-ztnsp_18f99d77-1c2d-412b-93ae-1d1fc52f24ab/machine-api-operator/0.log" Nov 23 07:31:55 crc kubenswrapper[4559]: E1123 07:31:55.919353 4559 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podffdd0197_ace4_42a8_9ccc_7a37ff23c2c3.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podffdd0197_ace4_42a8_9ccc_7a37ff23c2c3.slice/crio-4bf4376fb9b52f3e50bc811998b93b95334d953a90248d0578b5acfe08f9fe34\": RecentStats: unable to find data in memory cache]" Nov 23 07:31:56 crc kubenswrapper[4559]: I1123 07:31:56.167268 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:31:56 crc kubenswrapper[4559]: I1123 07:31:56.167324 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:31:56 crc kubenswrapper[4559]: I1123 07:31:56.302008 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-mqxbb_e5b5c7a9-2ccb-44ac-a6bd-826ffa8d7bda/cert-manager-controller/0.log" Nov 23 07:31:56 crc kubenswrapper[4559]: I1123 07:31:56.400701 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-m6h87_eb252f53-26dd-4071-8010-8a824926a679/cert-manager-cainjector/0.log" Nov 23 07:31:56 crc kubenswrapper[4559]: I1123 07:31:56.460626 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-qxxh5_74a9e3f9-a1b2-4f28-9ff4-a6d8a5d793b7/cert-manager-webhook/0.log" Nov 23 07:32:04 crc kubenswrapper[4559]: I1123 07:32:04.656365 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-6qvx9_b0662ab8-3035-4b7b-b615-15cc0b963af4/nmstate-console-plugin/0.log" Nov 23 07:32:04 crc kubenswrapper[4559]: I1123 07:32:04.778234 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-fglqj_092f86a7-8e24-4a70-9af6-a4265c3688e5/nmstate-handler/0.log" Nov 23 07:32:04 crc kubenswrapper[4559]: I1123 07:32:04.812654 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-rpd7c_3755e847-69fa-47e4-93f9-fe15df377011/nmstate-metrics/0.log" Nov 23 07:32:04 crc kubenswrapper[4559]: I1123 07:32:04.816263 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-rpd7c_3755e847-69fa-47e4-93f9-fe15df377011/kube-rbac-proxy/0.log" Nov 23 07:32:04 crc kubenswrapper[4559]: I1123 07:32:04.946654 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-pfk9b_c7cb504c-5b1d-4d00-b5e9-6bcc6968ee84/nmstate-operator/0.log" Nov 23 07:32:04 crc kubenswrapper[4559]: I1123 07:32:04.983750 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-hv2js_6ff08b99-2dc5-4b9f-bef5-b0d34e76cb49/nmstate-webhook/0.log" Nov 23 07:32:14 crc kubenswrapper[4559]: I1123 07:32:14.258906 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-tgxdj_24dffaef-1c72-43f9-a5a8-ee0797a63077/kube-rbac-proxy/0.log" Nov 23 07:32:14 crc kubenswrapper[4559]: I1123 07:32:14.364873 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-tgxdj_24dffaef-1c72-43f9-a5a8-ee0797a63077/controller/0.log" Nov 23 07:32:14 crc kubenswrapper[4559]: I1123 07:32:14.448837 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/cp-frr-files/0.log" Nov 23 07:32:14 crc kubenswrapper[4559]: I1123 07:32:14.535044 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/cp-frr-files/0.log" Nov 23 07:32:14 crc kubenswrapper[4559]: I1123 07:32:14.554509 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/cp-reloader/0.log" Nov 23 07:32:14 crc kubenswrapper[4559]: I1123 07:32:14.570936 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/cp-metrics/0.log" Nov 23 07:32:14 crc kubenswrapper[4559]: I1123 07:32:14.603320 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/cp-reloader/0.log" Nov 23 07:32:14 crc kubenswrapper[4559]: I1123 07:32:14.722961 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/cp-reloader/0.log" Nov 23 07:32:14 crc kubenswrapper[4559]: I1123 07:32:14.733540 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/cp-frr-files/0.log" Nov 23 07:32:14 crc kubenswrapper[4559]: I1123 07:32:14.761748 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/cp-metrics/0.log" Nov 23 07:32:14 crc kubenswrapper[4559]: I1123 07:32:14.765083 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/cp-metrics/0.log" Nov 23 07:32:14 crc kubenswrapper[4559]: I1123 07:32:14.994524 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/cp-frr-files/0.log" Nov 23 07:32:14 crc kubenswrapper[4559]: I1123 07:32:14.998175 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/cp-reloader/0.log" Nov 23 07:32:15 crc kubenswrapper[4559]: I1123 07:32:15.013699 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/cp-metrics/0.log" Nov 23 07:32:15 crc kubenswrapper[4559]: I1123 07:32:15.023175 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/controller/0.log" Nov 23 07:32:15 crc kubenswrapper[4559]: I1123 07:32:15.134531 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/frr-metrics/0.log" Nov 23 07:32:15 crc kubenswrapper[4559]: I1123 07:32:15.171784 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/kube-rbac-proxy-frr/0.log" Nov 23 07:32:15 crc kubenswrapper[4559]: I1123 07:32:15.238143 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/kube-rbac-proxy/0.log" Nov 23 07:32:15 crc kubenswrapper[4559]: I1123 07:32:15.304093 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/reloader/0.log" Nov 23 07:32:15 crc kubenswrapper[4559]: I1123 07:32:15.398349 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-zwlxs_5533445d-9129-4b9d-b87a-746df3caefb9/frr-k8s-webhook-server/0.log" Nov 23 07:32:15 crc kubenswrapper[4559]: I1123 07:32:15.539294 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5c8ccb79b7-ngqhc_e57be21d-6409-4c23-b745-ec26bdf7e98d/manager/0.log" Nov 23 07:32:15 crc kubenswrapper[4559]: I1123 07:32:15.678302 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6bc96cc64c-zfqmv_85db5798-c470-4ec9-91c3-c9bdd9db327d/webhook-server/0.log" Nov 23 07:32:15 crc kubenswrapper[4559]: I1123 07:32:15.743716 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-stnln_fe556537-f24f-4a84-9f79-488938b3f4b5/kube-rbac-proxy/0.log" Nov 23 07:32:16 crc kubenswrapper[4559]: I1123 07:32:16.103509 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/frr/0.log" Nov 23 07:32:16 crc kubenswrapper[4559]: I1123 07:32:16.152878 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-stnln_fe556537-f24f-4a84-9f79-488938b3f4b5/speaker/0.log" Nov 23 07:32:24 crc kubenswrapper[4559]: I1123 07:32:24.728465 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg_44c771c2-63aa-45f3-8551-f6b127f6121c/util/0.log" Nov 23 07:32:24 crc kubenswrapper[4559]: I1123 07:32:24.892320 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg_44c771c2-63aa-45f3-8551-f6b127f6121c/util/0.log" Nov 23 07:32:24 crc kubenswrapper[4559]: I1123 07:32:24.909705 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg_44c771c2-63aa-45f3-8551-f6b127f6121c/pull/0.log" Nov 23 07:32:24 crc kubenswrapper[4559]: I1123 07:32:24.916857 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg_44c771c2-63aa-45f3-8551-f6b127f6121c/pull/0.log" Nov 23 07:32:25 crc kubenswrapper[4559]: I1123 07:32:25.020069 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg_44c771c2-63aa-45f3-8551-f6b127f6121c/util/0.log" Nov 23 07:32:25 crc kubenswrapper[4559]: I1123 07:32:25.022213 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg_44c771c2-63aa-45f3-8551-f6b127f6121c/pull/0.log" Nov 23 07:32:25 crc kubenswrapper[4559]: I1123 07:32:25.030850 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg_44c771c2-63aa-45f3-8551-f6b127f6121c/extract/0.log" Nov 23 07:32:25 crc kubenswrapper[4559]: I1123 07:32:25.144906 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nc2rk_4d8f5bd0-c208-43fb-acd6-496d94f6dec3/extract-utilities/0.log" Nov 23 07:32:25 crc kubenswrapper[4559]: I1123 07:32:25.269815 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nc2rk_4d8f5bd0-c208-43fb-acd6-496d94f6dec3/extract-content/0.log" Nov 23 07:32:25 crc kubenswrapper[4559]: I1123 07:32:25.270955 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nc2rk_4d8f5bd0-c208-43fb-acd6-496d94f6dec3/extract-utilities/0.log" Nov 23 07:32:25 crc kubenswrapper[4559]: I1123 07:32:25.303290 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nc2rk_4d8f5bd0-c208-43fb-acd6-496d94f6dec3/extract-content/0.log" Nov 23 07:32:25 crc kubenswrapper[4559]: I1123 07:32:25.394155 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nc2rk_4d8f5bd0-c208-43fb-acd6-496d94f6dec3/extract-content/0.log" Nov 23 07:32:25 crc kubenswrapper[4559]: I1123 07:32:25.398657 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nc2rk_4d8f5bd0-c208-43fb-acd6-496d94f6dec3/extract-utilities/0.log" Nov 23 07:32:25 crc kubenswrapper[4559]: I1123 07:32:25.563930 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nw54h_c2daeb5e-241f-4764-962b-64c9c3d34597/extract-utilities/0.log" Nov 23 07:32:25 crc kubenswrapper[4559]: I1123 07:32:25.709237 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nc2rk_4d8f5bd0-c208-43fb-acd6-496d94f6dec3/registry-server/0.log" Nov 23 07:32:25 crc kubenswrapper[4559]: I1123 07:32:25.757630 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nw54h_c2daeb5e-241f-4764-962b-64c9c3d34597/extract-utilities/0.log" Nov 23 07:32:25 crc kubenswrapper[4559]: I1123 07:32:25.771904 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nw54h_c2daeb5e-241f-4764-962b-64c9c3d34597/extract-content/0.log" Nov 23 07:32:25 crc kubenswrapper[4559]: I1123 07:32:25.794333 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nw54h_c2daeb5e-241f-4764-962b-64c9c3d34597/extract-content/0.log" Nov 23 07:32:25 crc kubenswrapper[4559]: I1123 07:32:25.914849 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nw54h_c2daeb5e-241f-4764-962b-64c9c3d34597/extract-utilities/0.log" Nov 23 07:32:25 crc kubenswrapper[4559]: I1123 07:32:25.914894 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nw54h_c2daeb5e-241f-4764-962b-64c9c3d34597/extract-content/0.log" Nov 23 07:32:26 crc kubenswrapper[4559]: I1123 07:32:26.060003 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nw54h_c2daeb5e-241f-4764-962b-64c9c3d34597/registry-server/0.log" Nov 23 07:32:26 crc kubenswrapper[4559]: I1123 07:32:26.100549 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458_23f858ff-40d3-4f2a-ac00-f7e3b8c544f1/util/0.log" Nov 23 07:32:26 crc kubenswrapper[4559]: I1123 07:32:26.167433 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:32:26 crc kubenswrapper[4559]: I1123 07:32:26.167696 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:32:26 crc kubenswrapper[4559]: I1123 07:32:26.167806 4559 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" Nov 23 07:32:26 crc kubenswrapper[4559]: I1123 07:32:26.168486 4559 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4"} pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 07:32:26 crc kubenswrapper[4559]: I1123 07:32:26.168628 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" containerID="cri-o://c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4" gracePeriod=600 Nov 23 07:32:26 crc kubenswrapper[4559]: I1123 07:32:26.219719 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458_23f858ff-40d3-4f2a-ac00-f7e3b8c544f1/util/0.log" Nov 23 07:32:26 crc kubenswrapper[4559]: I1123 07:32:26.225048 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458_23f858ff-40d3-4f2a-ac00-f7e3b8c544f1/pull/0.log" Nov 23 07:32:26 crc kubenswrapper[4559]: I1123 07:32:26.272608 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458_23f858ff-40d3-4f2a-ac00-f7e3b8c544f1/pull/0.log" Nov 23 07:32:26 crc kubenswrapper[4559]: E1123 07:32:26.294126 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:32:26 crc kubenswrapper[4559]: I1123 07:32:26.356096 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458_23f858ff-40d3-4f2a-ac00-f7e3b8c544f1/extract/0.log" Nov 23 07:32:26 crc kubenswrapper[4559]: I1123 07:32:26.598009 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458_23f858ff-40d3-4f2a-ac00-f7e3b8c544f1/util/0.log" Nov 23 07:32:26 crc kubenswrapper[4559]: I1123 07:32:26.602336 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-v7hpt_351f8b04-cba3-4484-94b6-c4abd1c9be42/marketplace-operator/0.log" Nov 23 07:32:26 crc kubenswrapper[4559]: I1123 07:32:26.623123 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458_23f858ff-40d3-4f2a-ac00-f7e3b8c544f1/pull/0.log" Nov 23 07:32:26 crc kubenswrapper[4559]: I1123 07:32:26.732324 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4zrs4_b5a957ce-5a85-497c-9794-05ac9b6c92ae/extract-utilities/0.log" Nov 23 07:32:26 crc kubenswrapper[4559]: I1123 07:32:26.861564 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4zrs4_b5a957ce-5a85-497c-9794-05ac9b6c92ae/extract-content/0.log" Nov 23 07:32:26 crc kubenswrapper[4559]: I1123 07:32:26.865842 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4zrs4_b5a957ce-5a85-497c-9794-05ac9b6c92ae/extract-content/0.log" Nov 23 07:32:26 crc kubenswrapper[4559]: I1123 07:32:26.886466 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4zrs4_b5a957ce-5a85-497c-9794-05ac9b6c92ae/extract-utilities/0.log" Nov 23 07:32:27 crc kubenswrapper[4559]: I1123 07:32:27.010890 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4zrs4_b5a957ce-5a85-497c-9794-05ac9b6c92ae/extract-utilities/0.log" Nov 23 07:32:27 crc kubenswrapper[4559]: I1123 07:32:27.024195 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4zrs4_b5a957ce-5a85-497c-9794-05ac9b6c92ae/extract-content/0.log" Nov 23 07:32:27 crc kubenswrapper[4559]: I1123 07:32:27.105709 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4zrs4_b5a957ce-5a85-497c-9794-05ac9b6c92ae/registry-server/0.log" Nov 23 07:32:27 crc kubenswrapper[4559]: I1123 07:32:27.143176 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w2mts_0579529b-d4e5-458c-a2d6-9d0d7026c7b8/extract-utilities/0.log" Nov 23 07:32:27 crc kubenswrapper[4559]: I1123 07:32:27.143674 4559 generic.go:334] "Generic (PLEG): container finished" podID="4731beee-0cac-4189-8a70-743b0b709095" containerID="c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4" exitCode=0 Nov 23 07:32:27 crc kubenswrapper[4559]: I1123 07:32:27.143744 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerDied","Data":"c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4"} Nov 23 07:32:27 crc kubenswrapper[4559]: I1123 07:32:27.143874 4559 scope.go:117] "RemoveContainer" containerID="3b5b7bce47b314a27a0c4809c674e865bcfcc7e7b6194e23f9fb9dd9bd126d02" Nov 23 07:32:27 crc kubenswrapper[4559]: I1123 07:32:27.144917 4559 scope.go:117] "RemoveContainer" containerID="c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4" Nov 23 07:32:27 crc kubenswrapper[4559]: E1123 07:32:27.145392 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:32:27 crc kubenswrapper[4559]: I1123 07:32:27.319333 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w2mts_0579529b-d4e5-458c-a2d6-9d0d7026c7b8/extract-content/0.log" Nov 23 07:32:27 crc kubenswrapper[4559]: I1123 07:32:27.320235 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w2mts_0579529b-d4e5-458c-a2d6-9d0d7026c7b8/extract-utilities/0.log" Nov 23 07:32:27 crc kubenswrapper[4559]: I1123 07:32:27.333833 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w2mts_0579529b-d4e5-458c-a2d6-9d0d7026c7b8/extract-content/0.log" Nov 23 07:32:27 crc kubenswrapper[4559]: I1123 07:32:27.468930 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w2mts_0579529b-d4e5-458c-a2d6-9d0d7026c7b8/extract-content/0.log" Nov 23 07:32:27 crc kubenswrapper[4559]: I1123 07:32:27.470215 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w2mts_0579529b-d4e5-458c-a2d6-9d0d7026c7b8/extract-utilities/0.log" Nov 23 07:32:27 crc kubenswrapper[4559]: I1123 07:32:27.764342 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w2mts_0579529b-d4e5-458c-a2d6-9d0d7026c7b8/registry-server/0.log" Nov 23 07:32:39 crc kubenswrapper[4559]: I1123 07:32:39.273462 4559 scope.go:117] "RemoveContainer" containerID="c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4" Nov 23 07:32:39 crc kubenswrapper[4559]: E1123 07:32:39.275290 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:32:50 crc kubenswrapper[4559]: I1123 07:32:50.273347 4559 scope.go:117] "RemoveContainer" containerID="c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4" Nov 23 07:32:50 crc kubenswrapper[4559]: E1123 07:32:50.274001 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:32:50 crc kubenswrapper[4559]: I1123 07:32:50.743018 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-54wxf"] Nov 23 07:32:50 crc kubenswrapper[4559]: E1123 07:32:50.743625 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23ec849f-d94a-45b0-9f74-655f33ffb163" containerName="extract-utilities" Nov 23 07:32:50 crc kubenswrapper[4559]: I1123 07:32:50.743655 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="23ec849f-d94a-45b0-9f74-655f33ffb163" containerName="extract-utilities" Nov 23 07:32:50 crc kubenswrapper[4559]: E1123 07:32:50.743672 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3" containerName="container-00" Nov 23 07:32:50 crc kubenswrapper[4559]: I1123 07:32:50.743678 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3" containerName="container-00" Nov 23 07:32:50 crc kubenswrapper[4559]: E1123 07:32:50.743698 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23ec849f-d94a-45b0-9f74-655f33ffb163" containerName="registry-server" Nov 23 07:32:50 crc kubenswrapper[4559]: I1123 07:32:50.743704 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="23ec849f-d94a-45b0-9f74-655f33ffb163" containerName="registry-server" Nov 23 07:32:50 crc kubenswrapper[4559]: E1123 07:32:50.743722 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23ec849f-d94a-45b0-9f74-655f33ffb163" containerName="extract-content" Nov 23 07:32:50 crc kubenswrapper[4559]: I1123 07:32:50.743727 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="23ec849f-d94a-45b0-9f74-655f33ffb163" containerName="extract-content" Nov 23 07:32:50 crc kubenswrapper[4559]: I1123 07:32:50.743868 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffdd0197-ace4-42a8-9ccc-7a37ff23c2c3" containerName="container-00" Nov 23 07:32:50 crc kubenswrapper[4559]: I1123 07:32:50.743886 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="23ec849f-d94a-45b0-9f74-655f33ffb163" containerName="registry-server" Nov 23 07:32:50 crc kubenswrapper[4559]: I1123 07:32:50.745006 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-54wxf" Nov 23 07:32:50 crc kubenswrapper[4559]: I1123 07:32:50.754932 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-54wxf"] Nov 23 07:32:50 crc kubenswrapper[4559]: I1123 07:32:50.773625 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwrf5\" (UniqueName: \"kubernetes.io/projected/46b31641-6fb9-4c6b-95c1-6df261f7dcc3-kube-api-access-xwrf5\") pod \"redhat-marketplace-54wxf\" (UID: \"46b31641-6fb9-4c6b-95c1-6df261f7dcc3\") " pod="openshift-marketplace/redhat-marketplace-54wxf" Nov 23 07:32:50 crc kubenswrapper[4559]: I1123 07:32:50.773866 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46b31641-6fb9-4c6b-95c1-6df261f7dcc3-catalog-content\") pod \"redhat-marketplace-54wxf\" (UID: \"46b31641-6fb9-4c6b-95c1-6df261f7dcc3\") " pod="openshift-marketplace/redhat-marketplace-54wxf" Nov 23 07:32:50 crc kubenswrapper[4559]: I1123 07:32:50.773895 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46b31641-6fb9-4c6b-95c1-6df261f7dcc3-utilities\") pod \"redhat-marketplace-54wxf\" (UID: \"46b31641-6fb9-4c6b-95c1-6df261f7dcc3\") " pod="openshift-marketplace/redhat-marketplace-54wxf" Nov 23 07:32:50 crc kubenswrapper[4559]: I1123 07:32:50.875278 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46b31641-6fb9-4c6b-95c1-6df261f7dcc3-utilities\") pod \"redhat-marketplace-54wxf\" (UID: \"46b31641-6fb9-4c6b-95c1-6df261f7dcc3\") " pod="openshift-marketplace/redhat-marketplace-54wxf" Nov 23 07:32:50 crc kubenswrapper[4559]: I1123 07:32:50.875336 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwrf5\" (UniqueName: \"kubernetes.io/projected/46b31641-6fb9-4c6b-95c1-6df261f7dcc3-kube-api-access-xwrf5\") pod \"redhat-marketplace-54wxf\" (UID: \"46b31641-6fb9-4c6b-95c1-6df261f7dcc3\") " pod="openshift-marketplace/redhat-marketplace-54wxf" Nov 23 07:32:50 crc kubenswrapper[4559]: I1123 07:32:50.875449 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46b31641-6fb9-4c6b-95c1-6df261f7dcc3-catalog-content\") pod \"redhat-marketplace-54wxf\" (UID: \"46b31641-6fb9-4c6b-95c1-6df261f7dcc3\") " pod="openshift-marketplace/redhat-marketplace-54wxf" Nov 23 07:32:50 crc kubenswrapper[4559]: I1123 07:32:50.875928 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46b31641-6fb9-4c6b-95c1-6df261f7dcc3-catalog-content\") pod \"redhat-marketplace-54wxf\" (UID: \"46b31641-6fb9-4c6b-95c1-6df261f7dcc3\") " pod="openshift-marketplace/redhat-marketplace-54wxf" Nov 23 07:32:50 crc kubenswrapper[4559]: I1123 07:32:50.876034 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46b31641-6fb9-4c6b-95c1-6df261f7dcc3-utilities\") pod \"redhat-marketplace-54wxf\" (UID: \"46b31641-6fb9-4c6b-95c1-6df261f7dcc3\") " pod="openshift-marketplace/redhat-marketplace-54wxf" Nov 23 07:32:50 crc kubenswrapper[4559]: I1123 07:32:50.893103 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwrf5\" (UniqueName: \"kubernetes.io/projected/46b31641-6fb9-4c6b-95c1-6df261f7dcc3-kube-api-access-xwrf5\") pod \"redhat-marketplace-54wxf\" (UID: \"46b31641-6fb9-4c6b-95c1-6df261f7dcc3\") " pod="openshift-marketplace/redhat-marketplace-54wxf" Nov 23 07:32:51 crc kubenswrapper[4559]: I1123 07:32:51.069225 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-54wxf" Nov 23 07:32:51 crc kubenswrapper[4559]: I1123 07:32:51.475450 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-54wxf"] Nov 23 07:32:52 crc kubenswrapper[4559]: I1123 07:32:52.330943 4559 generic.go:334] "Generic (PLEG): container finished" podID="46b31641-6fb9-4c6b-95c1-6df261f7dcc3" containerID="4eb61f2f383be49c0af2cf8947a6a0db3203cc4bf551bc3a20e480da883af025" exitCode=0 Nov 23 07:32:52 crc kubenswrapper[4559]: I1123 07:32:52.331049 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-54wxf" event={"ID":"46b31641-6fb9-4c6b-95c1-6df261f7dcc3","Type":"ContainerDied","Data":"4eb61f2f383be49c0af2cf8947a6a0db3203cc4bf551bc3a20e480da883af025"} Nov 23 07:32:52 crc kubenswrapper[4559]: I1123 07:32:52.331305 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-54wxf" event={"ID":"46b31641-6fb9-4c6b-95c1-6df261f7dcc3","Type":"ContainerStarted","Data":"1fce2df1326566663491615e62dc5a4dc49cce31f064c64bdeec0fd84012bee4"} Nov 23 07:32:53 crc kubenswrapper[4559]: I1123 07:32:53.339101 4559 generic.go:334] "Generic (PLEG): container finished" podID="46b31641-6fb9-4c6b-95c1-6df261f7dcc3" containerID="b1124f6423d0fdc3ea70472dc1705a7f13548e61dabcef2168c80411fdec0e96" exitCode=0 Nov 23 07:32:53 crc kubenswrapper[4559]: I1123 07:32:53.339153 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-54wxf" event={"ID":"46b31641-6fb9-4c6b-95c1-6df261f7dcc3","Type":"ContainerDied","Data":"b1124f6423d0fdc3ea70472dc1705a7f13548e61dabcef2168c80411fdec0e96"} Nov 23 07:32:54 crc kubenswrapper[4559]: I1123 07:32:54.346842 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-54wxf" event={"ID":"46b31641-6fb9-4c6b-95c1-6df261f7dcc3","Type":"ContainerStarted","Data":"2b748db678ed564e8aa7c498d81a7f008160c10d0a5307277a2fc4482e4e7de1"} Nov 23 07:32:54 crc kubenswrapper[4559]: I1123 07:32:54.362287 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-54wxf" podStartSLOduration=2.876850901 podStartE2EDuration="4.362272829s" podCreationTimestamp="2025-11-23 07:32:50 +0000 UTC" firstStartedPulling="2025-11-23 07:32:52.333464647 +0000 UTC m=+2874.355450262" lastFinishedPulling="2025-11-23 07:32:53.818886585 +0000 UTC m=+2875.840872190" observedRunningTime="2025-11-23 07:32:54.360285253 +0000 UTC m=+2876.382270867" watchObservedRunningTime="2025-11-23 07:32:54.362272829 +0000 UTC m=+2876.384258433" Nov 23 07:33:01 crc kubenswrapper[4559]: I1123 07:33:01.069978 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-54wxf" Nov 23 07:33:01 crc kubenswrapper[4559]: I1123 07:33:01.070334 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-54wxf" Nov 23 07:33:01 crc kubenswrapper[4559]: I1123 07:33:01.101999 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-54wxf" Nov 23 07:33:01 crc kubenswrapper[4559]: I1123 07:33:01.421366 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-54wxf" Nov 23 07:33:01 crc kubenswrapper[4559]: I1123 07:33:01.455408 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-54wxf"] Nov 23 07:33:03 crc kubenswrapper[4559]: I1123 07:33:03.401939 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-54wxf" podUID="46b31641-6fb9-4c6b-95c1-6df261f7dcc3" containerName="registry-server" containerID="cri-o://2b748db678ed564e8aa7c498d81a7f008160c10d0a5307277a2fc4482e4e7de1" gracePeriod=2 Nov 23 07:33:03 crc kubenswrapper[4559]: I1123 07:33:03.762241 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-54wxf" Nov 23 07:33:03 crc kubenswrapper[4559]: I1123 07:33:03.954912 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46b31641-6fb9-4c6b-95c1-6df261f7dcc3-utilities\") pod \"46b31641-6fb9-4c6b-95c1-6df261f7dcc3\" (UID: \"46b31641-6fb9-4c6b-95c1-6df261f7dcc3\") " Nov 23 07:33:03 crc kubenswrapper[4559]: I1123 07:33:03.954960 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46b31641-6fb9-4c6b-95c1-6df261f7dcc3-catalog-content\") pod \"46b31641-6fb9-4c6b-95c1-6df261f7dcc3\" (UID: \"46b31641-6fb9-4c6b-95c1-6df261f7dcc3\") " Nov 23 07:33:03 crc kubenswrapper[4559]: I1123 07:33:03.954981 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwrf5\" (UniqueName: \"kubernetes.io/projected/46b31641-6fb9-4c6b-95c1-6df261f7dcc3-kube-api-access-xwrf5\") pod \"46b31641-6fb9-4c6b-95c1-6df261f7dcc3\" (UID: \"46b31641-6fb9-4c6b-95c1-6df261f7dcc3\") " Nov 23 07:33:03 crc kubenswrapper[4559]: I1123 07:33:03.956108 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46b31641-6fb9-4c6b-95c1-6df261f7dcc3-utilities" (OuterVolumeSpecName: "utilities") pod "46b31641-6fb9-4c6b-95c1-6df261f7dcc3" (UID: "46b31641-6fb9-4c6b-95c1-6df261f7dcc3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:33:03 crc kubenswrapper[4559]: I1123 07:33:03.962081 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46b31641-6fb9-4c6b-95c1-6df261f7dcc3-kube-api-access-xwrf5" (OuterVolumeSpecName: "kube-api-access-xwrf5") pod "46b31641-6fb9-4c6b-95c1-6df261f7dcc3" (UID: "46b31641-6fb9-4c6b-95c1-6df261f7dcc3"). InnerVolumeSpecName "kube-api-access-xwrf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:33:03 crc kubenswrapper[4559]: I1123 07:33:03.968741 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46b31641-6fb9-4c6b-95c1-6df261f7dcc3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "46b31641-6fb9-4c6b-95c1-6df261f7dcc3" (UID: "46b31641-6fb9-4c6b-95c1-6df261f7dcc3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:33:04 crc kubenswrapper[4559]: I1123 07:33:04.056210 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46b31641-6fb9-4c6b-95c1-6df261f7dcc3-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:33:04 crc kubenswrapper[4559]: I1123 07:33:04.056230 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46b31641-6fb9-4c6b-95c1-6df261f7dcc3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:33:04 crc kubenswrapper[4559]: I1123 07:33:04.056238 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwrf5\" (UniqueName: \"kubernetes.io/projected/46b31641-6fb9-4c6b-95c1-6df261f7dcc3-kube-api-access-xwrf5\") on node \"crc\" DevicePath \"\"" Nov 23 07:33:04 crc kubenswrapper[4559]: I1123 07:33:04.273265 4559 scope.go:117] "RemoveContainer" containerID="c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4" Nov 23 07:33:04 crc kubenswrapper[4559]: E1123 07:33:04.273549 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:33:04 crc kubenswrapper[4559]: I1123 07:33:04.410351 4559 generic.go:334] "Generic (PLEG): container finished" podID="46b31641-6fb9-4c6b-95c1-6df261f7dcc3" containerID="2b748db678ed564e8aa7c498d81a7f008160c10d0a5307277a2fc4482e4e7de1" exitCode=0 Nov 23 07:33:04 crc kubenswrapper[4559]: I1123 07:33:04.410391 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-54wxf" event={"ID":"46b31641-6fb9-4c6b-95c1-6df261f7dcc3","Type":"ContainerDied","Data":"2b748db678ed564e8aa7c498d81a7f008160c10d0a5307277a2fc4482e4e7de1"} Nov 23 07:33:04 crc kubenswrapper[4559]: I1123 07:33:04.410399 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-54wxf" Nov 23 07:33:04 crc kubenswrapper[4559]: I1123 07:33:04.410415 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-54wxf" event={"ID":"46b31641-6fb9-4c6b-95c1-6df261f7dcc3","Type":"ContainerDied","Data":"1fce2df1326566663491615e62dc5a4dc49cce31f064c64bdeec0fd84012bee4"} Nov 23 07:33:04 crc kubenswrapper[4559]: I1123 07:33:04.410432 4559 scope.go:117] "RemoveContainer" containerID="2b748db678ed564e8aa7c498d81a7f008160c10d0a5307277a2fc4482e4e7de1" Nov 23 07:33:04 crc kubenswrapper[4559]: I1123 07:33:04.425314 4559 scope.go:117] "RemoveContainer" containerID="b1124f6423d0fdc3ea70472dc1705a7f13548e61dabcef2168c80411fdec0e96" Nov 23 07:33:04 crc kubenswrapper[4559]: I1123 07:33:04.428028 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-54wxf"] Nov 23 07:33:04 crc kubenswrapper[4559]: I1123 07:33:04.434078 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-54wxf"] Nov 23 07:33:04 crc kubenswrapper[4559]: I1123 07:33:04.455589 4559 scope.go:117] "RemoveContainer" containerID="4eb61f2f383be49c0af2cf8947a6a0db3203cc4bf551bc3a20e480da883af025" Nov 23 07:33:04 crc kubenswrapper[4559]: I1123 07:33:04.473166 4559 scope.go:117] "RemoveContainer" containerID="2b748db678ed564e8aa7c498d81a7f008160c10d0a5307277a2fc4482e4e7de1" Nov 23 07:33:04 crc kubenswrapper[4559]: E1123 07:33:04.473518 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b748db678ed564e8aa7c498d81a7f008160c10d0a5307277a2fc4482e4e7de1\": container with ID starting with 2b748db678ed564e8aa7c498d81a7f008160c10d0a5307277a2fc4482e4e7de1 not found: ID does not exist" containerID="2b748db678ed564e8aa7c498d81a7f008160c10d0a5307277a2fc4482e4e7de1" Nov 23 07:33:04 crc kubenswrapper[4559]: I1123 07:33:04.473609 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b748db678ed564e8aa7c498d81a7f008160c10d0a5307277a2fc4482e4e7de1"} err="failed to get container status \"2b748db678ed564e8aa7c498d81a7f008160c10d0a5307277a2fc4482e4e7de1\": rpc error: code = NotFound desc = could not find container \"2b748db678ed564e8aa7c498d81a7f008160c10d0a5307277a2fc4482e4e7de1\": container with ID starting with 2b748db678ed564e8aa7c498d81a7f008160c10d0a5307277a2fc4482e4e7de1 not found: ID does not exist" Nov 23 07:33:04 crc kubenswrapper[4559]: I1123 07:33:04.473701 4559 scope.go:117] "RemoveContainer" containerID="b1124f6423d0fdc3ea70472dc1705a7f13548e61dabcef2168c80411fdec0e96" Nov 23 07:33:04 crc kubenswrapper[4559]: E1123 07:33:04.473986 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1124f6423d0fdc3ea70472dc1705a7f13548e61dabcef2168c80411fdec0e96\": container with ID starting with b1124f6423d0fdc3ea70472dc1705a7f13548e61dabcef2168c80411fdec0e96 not found: ID does not exist" containerID="b1124f6423d0fdc3ea70472dc1705a7f13548e61dabcef2168c80411fdec0e96" Nov 23 07:33:04 crc kubenswrapper[4559]: I1123 07:33:04.474064 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1124f6423d0fdc3ea70472dc1705a7f13548e61dabcef2168c80411fdec0e96"} err="failed to get container status \"b1124f6423d0fdc3ea70472dc1705a7f13548e61dabcef2168c80411fdec0e96\": rpc error: code = NotFound desc = could not find container \"b1124f6423d0fdc3ea70472dc1705a7f13548e61dabcef2168c80411fdec0e96\": container with ID starting with b1124f6423d0fdc3ea70472dc1705a7f13548e61dabcef2168c80411fdec0e96 not found: ID does not exist" Nov 23 07:33:04 crc kubenswrapper[4559]: I1123 07:33:04.474145 4559 scope.go:117] "RemoveContainer" containerID="4eb61f2f383be49c0af2cf8947a6a0db3203cc4bf551bc3a20e480da883af025" Nov 23 07:33:04 crc kubenswrapper[4559]: E1123 07:33:04.474441 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4eb61f2f383be49c0af2cf8947a6a0db3203cc4bf551bc3a20e480da883af025\": container with ID starting with 4eb61f2f383be49c0af2cf8947a6a0db3203cc4bf551bc3a20e480da883af025 not found: ID does not exist" containerID="4eb61f2f383be49c0af2cf8947a6a0db3203cc4bf551bc3a20e480da883af025" Nov 23 07:33:04 crc kubenswrapper[4559]: I1123 07:33:04.474463 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4eb61f2f383be49c0af2cf8947a6a0db3203cc4bf551bc3a20e480da883af025"} err="failed to get container status \"4eb61f2f383be49c0af2cf8947a6a0db3203cc4bf551bc3a20e480da883af025\": rpc error: code = NotFound desc = could not find container \"4eb61f2f383be49c0af2cf8947a6a0db3203cc4bf551bc3a20e480da883af025\": container with ID starting with 4eb61f2f383be49c0af2cf8947a6a0db3203cc4bf551bc3a20e480da883af025 not found: ID does not exist" Nov 23 07:33:06 crc kubenswrapper[4559]: I1123 07:33:06.281180 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46b31641-6fb9-4c6b-95c1-6df261f7dcc3" path="/var/lib/kubelet/pods/46b31641-6fb9-4c6b-95c1-6df261f7dcc3/volumes" Nov 23 07:33:17 crc kubenswrapper[4559]: I1123 07:33:17.274029 4559 scope.go:117] "RemoveContainer" containerID="c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4" Nov 23 07:33:17 crc kubenswrapper[4559]: E1123 07:33:17.274846 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:33:30 crc kubenswrapper[4559]: I1123 07:33:30.274048 4559 scope.go:117] "RemoveContainer" containerID="c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4" Nov 23 07:33:30 crc kubenswrapper[4559]: E1123 07:33:30.274674 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:33:39 crc kubenswrapper[4559]: I1123 07:33:39.631990 4559 generic.go:334] "Generic (PLEG): container finished" podID="13525b40-ea0a-4894-8622-e017b4f34397" containerID="956585f1ea41a294bbff57232074686727d455815232ba5506a96bcfde6a105a" exitCode=0 Nov 23 07:33:39 crc kubenswrapper[4559]: I1123 07:33:39.632078 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m8wzp/must-gather-8nhbp" event={"ID":"13525b40-ea0a-4894-8622-e017b4f34397","Type":"ContainerDied","Data":"956585f1ea41a294bbff57232074686727d455815232ba5506a96bcfde6a105a"} Nov 23 07:33:39 crc kubenswrapper[4559]: I1123 07:33:39.632936 4559 scope.go:117] "RemoveContainer" containerID="956585f1ea41a294bbff57232074686727d455815232ba5506a96bcfde6a105a" Nov 23 07:33:39 crc kubenswrapper[4559]: I1123 07:33:39.798557 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-m8wzp_must-gather-8nhbp_13525b40-ea0a-4894-8622-e017b4f34397/gather/0.log" Nov 23 07:33:42 crc kubenswrapper[4559]: I1123 07:33:42.273505 4559 scope.go:117] "RemoveContainer" containerID="c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4" Nov 23 07:33:42 crc kubenswrapper[4559]: E1123 07:33:42.273743 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:33:46 crc kubenswrapper[4559]: I1123 07:33:46.449505 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-m8wzp/must-gather-8nhbp"] Nov 23 07:33:46 crc kubenswrapper[4559]: I1123 07:33:46.449911 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-m8wzp/must-gather-8nhbp" podUID="13525b40-ea0a-4894-8622-e017b4f34397" containerName="copy" containerID="cri-o://5499865a9a9dd4b62b7fc407abdf7e24b86e7fee32f9b16260a32c7058847885" gracePeriod=2 Nov 23 07:33:46 crc kubenswrapper[4559]: I1123 07:33:46.455374 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-m8wzp/must-gather-8nhbp"] Nov 23 07:33:46 crc kubenswrapper[4559]: I1123 07:33:46.676357 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-m8wzp_must-gather-8nhbp_13525b40-ea0a-4894-8622-e017b4f34397/copy/0.log" Nov 23 07:33:46 crc kubenswrapper[4559]: I1123 07:33:46.676864 4559 generic.go:334] "Generic (PLEG): container finished" podID="13525b40-ea0a-4894-8622-e017b4f34397" containerID="5499865a9a9dd4b62b7fc407abdf7e24b86e7fee32f9b16260a32c7058847885" exitCode=143 Nov 23 07:33:46 crc kubenswrapper[4559]: I1123 07:33:46.788397 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-m8wzp_must-gather-8nhbp_13525b40-ea0a-4894-8622-e017b4f34397/copy/0.log" Nov 23 07:33:46 crc kubenswrapper[4559]: I1123 07:33:46.789014 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m8wzp/must-gather-8nhbp" Nov 23 07:33:46 crc kubenswrapper[4559]: I1123 07:33:46.918276 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/13525b40-ea0a-4894-8622-e017b4f34397-must-gather-output\") pod \"13525b40-ea0a-4894-8622-e017b4f34397\" (UID: \"13525b40-ea0a-4894-8622-e017b4f34397\") " Nov 23 07:33:46 crc kubenswrapper[4559]: I1123 07:33:46.918411 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6b8z\" (UniqueName: \"kubernetes.io/projected/13525b40-ea0a-4894-8622-e017b4f34397-kube-api-access-q6b8z\") pod \"13525b40-ea0a-4894-8622-e017b4f34397\" (UID: \"13525b40-ea0a-4894-8622-e017b4f34397\") " Nov 23 07:33:46 crc kubenswrapper[4559]: I1123 07:33:46.923923 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13525b40-ea0a-4894-8622-e017b4f34397-kube-api-access-q6b8z" (OuterVolumeSpecName: "kube-api-access-q6b8z") pod "13525b40-ea0a-4894-8622-e017b4f34397" (UID: "13525b40-ea0a-4894-8622-e017b4f34397"). InnerVolumeSpecName "kube-api-access-q6b8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:33:47 crc kubenswrapper[4559]: I1123 07:33:47.021024 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6b8z\" (UniqueName: \"kubernetes.io/projected/13525b40-ea0a-4894-8622-e017b4f34397-kube-api-access-q6b8z\") on node \"crc\" DevicePath \"\"" Nov 23 07:33:47 crc kubenswrapper[4559]: I1123 07:33:47.025735 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13525b40-ea0a-4894-8622-e017b4f34397-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "13525b40-ea0a-4894-8622-e017b4f34397" (UID: "13525b40-ea0a-4894-8622-e017b4f34397"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:33:47 crc kubenswrapper[4559]: I1123 07:33:47.122358 4559 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/13525b40-ea0a-4894-8622-e017b4f34397-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 23 07:33:47 crc kubenswrapper[4559]: I1123 07:33:47.685192 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-m8wzp_must-gather-8nhbp_13525b40-ea0a-4894-8622-e017b4f34397/copy/0.log" Nov 23 07:33:47 crc kubenswrapper[4559]: I1123 07:33:47.685599 4559 scope.go:117] "RemoveContainer" containerID="5499865a9a9dd4b62b7fc407abdf7e24b86e7fee32f9b16260a32c7058847885" Nov 23 07:33:47 crc kubenswrapper[4559]: I1123 07:33:47.685638 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m8wzp/must-gather-8nhbp" Nov 23 07:33:47 crc kubenswrapper[4559]: I1123 07:33:47.701166 4559 scope.go:117] "RemoveContainer" containerID="956585f1ea41a294bbff57232074686727d455815232ba5506a96bcfde6a105a" Nov 23 07:33:48 crc kubenswrapper[4559]: I1123 07:33:48.281839 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13525b40-ea0a-4894-8622-e017b4f34397" path="/var/lib/kubelet/pods/13525b40-ea0a-4894-8622-e017b4f34397/volumes" Nov 23 07:33:53 crc kubenswrapper[4559]: I1123 07:33:53.274065 4559 scope.go:117] "RemoveContainer" containerID="c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4" Nov 23 07:33:53 crc kubenswrapper[4559]: E1123 07:33:53.274592 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:34:07 crc kubenswrapper[4559]: I1123 07:34:07.273396 4559 scope.go:117] "RemoveContainer" containerID="c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4" Nov 23 07:34:07 crc kubenswrapper[4559]: E1123 07:34:07.273951 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:34:22 crc kubenswrapper[4559]: I1123 07:34:22.273388 4559 scope.go:117] "RemoveContainer" containerID="c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4" Nov 23 07:34:22 crc kubenswrapper[4559]: E1123 07:34:22.273972 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:34:35 crc kubenswrapper[4559]: I1123 07:34:35.273392 4559 scope.go:117] "RemoveContainer" containerID="c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4" Nov 23 07:34:35 crc kubenswrapper[4559]: E1123 07:34:35.273930 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:34:47 crc kubenswrapper[4559]: I1123 07:34:47.273271 4559 scope.go:117] "RemoveContainer" containerID="c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4" Nov 23 07:34:47 crc kubenswrapper[4559]: E1123 07:34:47.273909 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:34:58 crc kubenswrapper[4559]: I1123 07:34:58.280548 4559 scope.go:117] "RemoveContainer" containerID="c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4" Nov 23 07:34:58 crc kubenswrapper[4559]: E1123 07:34:58.281393 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:35:11 crc kubenswrapper[4559]: I1123 07:35:11.273813 4559 scope.go:117] "RemoveContainer" containerID="c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4" Nov 23 07:35:11 crc kubenswrapper[4559]: E1123 07:35:11.274433 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:35:24 crc kubenswrapper[4559]: I1123 07:35:24.274150 4559 scope.go:117] "RemoveContainer" containerID="c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4" Nov 23 07:35:24 crc kubenswrapper[4559]: E1123 07:35:24.274718 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:35:35 crc kubenswrapper[4559]: I1123 07:35:35.273921 4559 scope.go:117] "RemoveContainer" containerID="c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4" Nov 23 07:35:35 crc kubenswrapper[4559]: E1123 07:35:35.274435 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:35:42 crc kubenswrapper[4559]: I1123 07:35:42.137293 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-m5msz/must-gather-fgqp7"] Nov 23 07:35:42 crc kubenswrapper[4559]: E1123 07:35:42.137934 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46b31641-6fb9-4c6b-95c1-6df261f7dcc3" containerName="registry-server" Nov 23 07:35:42 crc kubenswrapper[4559]: I1123 07:35:42.137947 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="46b31641-6fb9-4c6b-95c1-6df261f7dcc3" containerName="registry-server" Nov 23 07:35:42 crc kubenswrapper[4559]: E1123 07:35:42.137965 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46b31641-6fb9-4c6b-95c1-6df261f7dcc3" containerName="extract-utilities" Nov 23 07:35:42 crc kubenswrapper[4559]: I1123 07:35:42.137971 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="46b31641-6fb9-4c6b-95c1-6df261f7dcc3" containerName="extract-utilities" Nov 23 07:35:42 crc kubenswrapper[4559]: E1123 07:35:42.137988 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46b31641-6fb9-4c6b-95c1-6df261f7dcc3" containerName="extract-content" Nov 23 07:35:42 crc kubenswrapper[4559]: I1123 07:35:42.137994 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="46b31641-6fb9-4c6b-95c1-6df261f7dcc3" containerName="extract-content" Nov 23 07:35:42 crc kubenswrapper[4559]: E1123 07:35:42.138001 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13525b40-ea0a-4894-8622-e017b4f34397" containerName="gather" Nov 23 07:35:42 crc kubenswrapper[4559]: I1123 07:35:42.138006 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="13525b40-ea0a-4894-8622-e017b4f34397" containerName="gather" Nov 23 07:35:42 crc kubenswrapper[4559]: E1123 07:35:42.138022 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13525b40-ea0a-4894-8622-e017b4f34397" containerName="copy" Nov 23 07:35:42 crc kubenswrapper[4559]: I1123 07:35:42.138028 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="13525b40-ea0a-4894-8622-e017b4f34397" containerName="copy" Nov 23 07:35:42 crc kubenswrapper[4559]: I1123 07:35:42.138212 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="46b31641-6fb9-4c6b-95c1-6df261f7dcc3" containerName="registry-server" Nov 23 07:35:42 crc kubenswrapper[4559]: I1123 07:35:42.138223 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="13525b40-ea0a-4894-8622-e017b4f34397" containerName="gather" Nov 23 07:35:42 crc kubenswrapper[4559]: I1123 07:35:42.138234 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="13525b40-ea0a-4894-8622-e017b4f34397" containerName="copy" Nov 23 07:35:42 crc kubenswrapper[4559]: I1123 07:35:42.139057 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m5msz/must-gather-fgqp7" Nov 23 07:35:42 crc kubenswrapper[4559]: I1123 07:35:42.141393 4559 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-m5msz"/"default-dockercfg-j7g7w" Nov 23 07:35:42 crc kubenswrapper[4559]: I1123 07:35:42.141769 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-m5msz"/"openshift-service-ca.crt" Nov 23 07:35:42 crc kubenswrapper[4559]: I1123 07:35:42.141898 4559 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-m5msz"/"kube-root-ca.crt" Nov 23 07:35:42 crc kubenswrapper[4559]: I1123 07:35:42.156370 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-m5msz/must-gather-fgqp7"] Nov 23 07:35:42 crc kubenswrapper[4559]: I1123 07:35:42.198485 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbxts\" (UniqueName: \"kubernetes.io/projected/08ac5460-3a47-487d-9b86-93c6debd86d9-kube-api-access-gbxts\") pod \"must-gather-fgqp7\" (UID: \"08ac5460-3a47-487d-9b86-93c6debd86d9\") " pod="openshift-must-gather-m5msz/must-gather-fgqp7" Nov 23 07:35:42 crc kubenswrapper[4559]: I1123 07:35:42.198572 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/08ac5460-3a47-487d-9b86-93c6debd86d9-must-gather-output\") pod \"must-gather-fgqp7\" (UID: \"08ac5460-3a47-487d-9b86-93c6debd86d9\") " pod="openshift-must-gather-m5msz/must-gather-fgqp7" Nov 23 07:35:42 crc kubenswrapper[4559]: I1123 07:35:42.300896 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbxts\" (UniqueName: \"kubernetes.io/projected/08ac5460-3a47-487d-9b86-93c6debd86d9-kube-api-access-gbxts\") pod \"must-gather-fgqp7\" (UID: \"08ac5460-3a47-487d-9b86-93c6debd86d9\") " pod="openshift-must-gather-m5msz/must-gather-fgqp7" Nov 23 07:35:42 crc kubenswrapper[4559]: I1123 07:35:42.301195 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/08ac5460-3a47-487d-9b86-93c6debd86d9-must-gather-output\") pod \"must-gather-fgqp7\" (UID: \"08ac5460-3a47-487d-9b86-93c6debd86d9\") " pod="openshift-must-gather-m5msz/must-gather-fgqp7" Nov 23 07:35:42 crc kubenswrapper[4559]: I1123 07:35:42.301546 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/08ac5460-3a47-487d-9b86-93c6debd86d9-must-gather-output\") pod \"must-gather-fgqp7\" (UID: \"08ac5460-3a47-487d-9b86-93c6debd86d9\") " pod="openshift-must-gather-m5msz/must-gather-fgqp7" Nov 23 07:35:42 crc kubenswrapper[4559]: I1123 07:35:42.315484 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbxts\" (UniqueName: \"kubernetes.io/projected/08ac5460-3a47-487d-9b86-93c6debd86d9-kube-api-access-gbxts\") pod \"must-gather-fgqp7\" (UID: \"08ac5460-3a47-487d-9b86-93c6debd86d9\") " pod="openshift-must-gather-m5msz/must-gather-fgqp7" Nov 23 07:35:42 crc kubenswrapper[4559]: I1123 07:35:42.454915 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m5msz/must-gather-fgqp7" Nov 23 07:35:42 crc kubenswrapper[4559]: I1123 07:35:42.710987 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-m5msz/must-gather-fgqp7"] Nov 23 07:35:43 crc kubenswrapper[4559]: I1123 07:35:43.434658 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m5msz/must-gather-fgqp7" event={"ID":"08ac5460-3a47-487d-9b86-93c6debd86d9","Type":"ContainerStarted","Data":"abeb131c06d46e0814ddc07e00b54f898ca86cf245e8912002af75411617152c"} Nov 23 07:35:43 crc kubenswrapper[4559]: I1123 07:35:43.434854 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m5msz/must-gather-fgqp7" event={"ID":"08ac5460-3a47-487d-9b86-93c6debd86d9","Type":"ContainerStarted","Data":"7b6a158235ebef33f14538cf8e75beb0fadadd55a9ef4c3fb415fa7c9a93e5a2"} Nov 23 07:35:43 crc kubenswrapper[4559]: I1123 07:35:43.434866 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m5msz/must-gather-fgqp7" event={"ID":"08ac5460-3a47-487d-9b86-93c6debd86d9","Type":"ContainerStarted","Data":"46d65f5a0bc953fcfe670b0fd3e35a5973705f058d65d6fa230bb4853b4aaacb"} Nov 23 07:35:43 crc kubenswrapper[4559]: I1123 07:35:43.449865 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-m5msz/must-gather-fgqp7" podStartSLOduration=1.449855178 podStartE2EDuration="1.449855178s" podCreationTimestamp="2025-11-23 07:35:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:35:43.448172856 +0000 UTC m=+3045.470158490" watchObservedRunningTime="2025-11-23 07:35:43.449855178 +0000 UTC m=+3045.471840792" Nov 23 07:35:45 crc kubenswrapper[4559]: I1123 07:35:45.392001 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-m5msz/crc-debug-426z9"] Nov 23 07:35:45 crc kubenswrapper[4559]: I1123 07:35:45.393385 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m5msz/crc-debug-426z9" Nov 23 07:35:45 crc kubenswrapper[4559]: I1123 07:35:45.452296 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5nk9\" (UniqueName: \"kubernetes.io/projected/51c39ef5-3bee-4998-b67c-f9f51aa5ec48-kube-api-access-m5nk9\") pod \"crc-debug-426z9\" (UID: \"51c39ef5-3bee-4998-b67c-f9f51aa5ec48\") " pod="openshift-must-gather-m5msz/crc-debug-426z9" Nov 23 07:35:45 crc kubenswrapper[4559]: I1123 07:35:45.452680 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/51c39ef5-3bee-4998-b67c-f9f51aa5ec48-host\") pod \"crc-debug-426z9\" (UID: \"51c39ef5-3bee-4998-b67c-f9f51aa5ec48\") " pod="openshift-must-gather-m5msz/crc-debug-426z9" Nov 23 07:35:45 crc kubenswrapper[4559]: I1123 07:35:45.554156 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/51c39ef5-3bee-4998-b67c-f9f51aa5ec48-host\") pod \"crc-debug-426z9\" (UID: \"51c39ef5-3bee-4998-b67c-f9f51aa5ec48\") " pod="openshift-must-gather-m5msz/crc-debug-426z9" Nov 23 07:35:45 crc kubenswrapper[4559]: I1123 07:35:45.554212 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5nk9\" (UniqueName: \"kubernetes.io/projected/51c39ef5-3bee-4998-b67c-f9f51aa5ec48-kube-api-access-m5nk9\") pod \"crc-debug-426z9\" (UID: \"51c39ef5-3bee-4998-b67c-f9f51aa5ec48\") " pod="openshift-must-gather-m5msz/crc-debug-426z9" Nov 23 07:35:45 crc kubenswrapper[4559]: I1123 07:35:45.554278 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/51c39ef5-3bee-4998-b67c-f9f51aa5ec48-host\") pod \"crc-debug-426z9\" (UID: \"51c39ef5-3bee-4998-b67c-f9f51aa5ec48\") " pod="openshift-must-gather-m5msz/crc-debug-426z9" Nov 23 07:35:45 crc kubenswrapper[4559]: I1123 07:35:45.570820 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5nk9\" (UniqueName: \"kubernetes.io/projected/51c39ef5-3bee-4998-b67c-f9f51aa5ec48-kube-api-access-m5nk9\") pod \"crc-debug-426z9\" (UID: \"51c39ef5-3bee-4998-b67c-f9f51aa5ec48\") " pod="openshift-must-gather-m5msz/crc-debug-426z9" Nov 23 07:35:45 crc kubenswrapper[4559]: I1123 07:35:45.709484 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m5msz/crc-debug-426z9" Nov 23 07:35:45 crc kubenswrapper[4559]: W1123 07:35:45.773563 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51c39ef5_3bee_4998_b67c_f9f51aa5ec48.slice/crio-ea17030debcc5a1bb84f44bd29a4c046758b37a09d025740922e85f8dd768dae WatchSource:0}: Error finding container ea17030debcc5a1bb84f44bd29a4c046758b37a09d025740922e85f8dd768dae: Status 404 returned error can't find the container with id ea17030debcc5a1bb84f44bd29a4c046758b37a09d025740922e85f8dd768dae Nov 23 07:35:46 crc kubenswrapper[4559]: I1123 07:35:46.273285 4559 scope.go:117] "RemoveContainer" containerID="c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4" Nov 23 07:35:46 crc kubenswrapper[4559]: E1123 07:35:46.273995 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:35:46 crc kubenswrapper[4559]: I1123 07:35:46.480000 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m5msz/crc-debug-426z9" event={"ID":"51c39ef5-3bee-4998-b67c-f9f51aa5ec48","Type":"ContainerStarted","Data":"c2f7210eb3f838b99150da4d583427f03f082a736f9a65fbfe42d31a2659d3a9"} Nov 23 07:35:46 crc kubenswrapper[4559]: I1123 07:35:46.480052 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m5msz/crc-debug-426z9" event={"ID":"51c39ef5-3bee-4998-b67c-f9f51aa5ec48","Type":"ContainerStarted","Data":"ea17030debcc5a1bb84f44bd29a4c046758b37a09d025740922e85f8dd768dae"} Nov 23 07:35:46 crc kubenswrapper[4559]: I1123 07:35:46.495202 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-m5msz/crc-debug-426z9" podStartSLOduration=1.495187723 podStartE2EDuration="1.495187723s" podCreationTimestamp="2025-11-23 07:35:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 07:35:46.488693705 +0000 UTC m=+3048.510679320" watchObservedRunningTime="2025-11-23 07:35:46.495187723 +0000 UTC m=+3048.517173337" Nov 23 07:35:57 crc kubenswrapper[4559]: I1123 07:35:57.274402 4559 scope.go:117] "RemoveContainer" containerID="c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4" Nov 23 07:35:57 crc kubenswrapper[4559]: E1123 07:35:57.275052 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:35:59 crc kubenswrapper[4559]: I1123 07:35:59.848349 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jflsv"] Nov 23 07:35:59 crc kubenswrapper[4559]: I1123 07:35:59.850325 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jflsv" Nov 23 07:35:59 crc kubenswrapper[4559]: I1123 07:35:59.857056 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jflsv"] Nov 23 07:35:59 crc kubenswrapper[4559]: I1123 07:35:59.878716 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cb21f8e-86bb-48f1-9510-510ad0a78890-catalog-content\") pod \"redhat-operators-jflsv\" (UID: \"0cb21f8e-86bb-48f1-9510-510ad0a78890\") " pod="openshift-marketplace/redhat-operators-jflsv" Nov 23 07:35:59 crc kubenswrapper[4559]: I1123 07:35:59.878899 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cb21f8e-86bb-48f1-9510-510ad0a78890-utilities\") pod \"redhat-operators-jflsv\" (UID: \"0cb21f8e-86bb-48f1-9510-510ad0a78890\") " pod="openshift-marketplace/redhat-operators-jflsv" Nov 23 07:35:59 crc kubenswrapper[4559]: I1123 07:35:59.878985 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4zc6\" (UniqueName: \"kubernetes.io/projected/0cb21f8e-86bb-48f1-9510-510ad0a78890-kube-api-access-z4zc6\") pod \"redhat-operators-jflsv\" (UID: \"0cb21f8e-86bb-48f1-9510-510ad0a78890\") " pod="openshift-marketplace/redhat-operators-jflsv" Nov 23 07:35:59 crc kubenswrapper[4559]: I1123 07:35:59.980827 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4zc6\" (UniqueName: \"kubernetes.io/projected/0cb21f8e-86bb-48f1-9510-510ad0a78890-kube-api-access-z4zc6\") pod \"redhat-operators-jflsv\" (UID: \"0cb21f8e-86bb-48f1-9510-510ad0a78890\") " pod="openshift-marketplace/redhat-operators-jflsv" Nov 23 07:35:59 crc kubenswrapper[4559]: I1123 07:35:59.980980 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cb21f8e-86bb-48f1-9510-510ad0a78890-catalog-content\") pod \"redhat-operators-jflsv\" (UID: \"0cb21f8e-86bb-48f1-9510-510ad0a78890\") " pod="openshift-marketplace/redhat-operators-jflsv" Nov 23 07:35:59 crc kubenswrapper[4559]: I1123 07:35:59.981115 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cb21f8e-86bb-48f1-9510-510ad0a78890-utilities\") pod \"redhat-operators-jflsv\" (UID: \"0cb21f8e-86bb-48f1-9510-510ad0a78890\") " pod="openshift-marketplace/redhat-operators-jflsv" Nov 23 07:35:59 crc kubenswrapper[4559]: I1123 07:35:59.981374 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cb21f8e-86bb-48f1-9510-510ad0a78890-catalog-content\") pod \"redhat-operators-jflsv\" (UID: \"0cb21f8e-86bb-48f1-9510-510ad0a78890\") " pod="openshift-marketplace/redhat-operators-jflsv" Nov 23 07:35:59 crc kubenswrapper[4559]: I1123 07:35:59.981433 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cb21f8e-86bb-48f1-9510-510ad0a78890-utilities\") pod \"redhat-operators-jflsv\" (UID: \"0cb21f8e-86bb-48f1-9510-510ad0a78890\") " pod="openshift-marketplace/redhat-operators-jflsv" Nov 23 07:36:00 crc kubenswrapper[4559]: I1123 07:36:00.002158 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4zc6\" (UniqueName: \"kubernetes.io/projected/0cb21f8e-86bb-48f1-9510-510ad0a78890-kube-api-access-z4zc6\") pod \"redhat-operators-jflsv\" (UID: \"0cb21f8e-86bb-48f1-9510-510ad0a78890\") " pod="openshift-marketplace/redhat-operators-jflsv" Nov 23 07:36:00 crc kubenswrapper[4559]: I1123 07:36:00.164197 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jflsv" Nov 23 07:36:00 crc kubenswrapper[4559]: I1123 07:36:00.660892 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jflsv"] Nov 23 07:36:01 crc kubenswrapper[4559]: I1123 07:36:01.586953 4559 generic.go:334] "Generic (PLEG): container finished" podID="0cb21f8e-86bb-48f1-9510-510ad0a78890" containerID="f4321a1001380e6fe72d86292078c37a1eac98137c8bdc9165dd64e3a482443b" exitCode=0 Nov 23 07:36:01 crc kubenswrapper[4559]: I1123 07:36:01.587050 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jflsv" event={"ID":"0cb21f8e-86bb-48f1-9510-510ad0a78890","Type":"ContainerDied","Data":"f4321a1001380e6fe72d86292078c37a1eac98137c8bdc9165dd64e3a482443b"} Nov 23 07:36:01 crc kubenswrapper[4559]: I1123 07:36:01.587205 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jflsv" event={"ID":"0cb21f8e-86bb-48f1-9510-510ad0a78890","Type":"ContainerStarted","Data":"cacce27b257e7f5afc70eaa69665d0c774ad78989946bb06931d228f6686755a"} Nov 23 07:36:01 crc kubenswrapper[4559]: I1123 07:36:01.588725 4559 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 07:36:02 crc kubenswrapper[4559]: I1123 07:36:02.594338 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jflsv" event={"ID":"0cb21f8e-86bb-48f1-9510-510ad0a78890","Type":"ContainerStarted","Data":"9f531680a4968a698989537855c95acb162cd99751a3d1c3ab35605d0e818ce5"} Nov 23 07:36:03 crc kubenswrapper[4559]: I1123 07:36:03.602180 4559 generic.go:334] "Generic (PLEG): container finished" podID="0cb21f8e-86bb-48f1-9510-510ad0a78890" containerID="9f531680a4968a698989537855c95acb162cd99751a3d1c3ab35605d0e818ce5" exitCode=0 Nov 23 07:36:03 crc kubenswrapper[4559]: I1123 07:36:03.602276 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jflsv" event={"ID":"0cb21f8e-86bb-48f1-9510-510ad0a78890","Type":"ContainerDied","Data":"9f531680a4968a698989537855c95acb162cd99751a3d1c3ab35605d0e818ce5"} Nov 23 07:36:04 crc kubenswrapper[4559]: I1123 07:36:04.611336 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jflsv" event={"ID":"0cb21f8e-86bb-48f1-9510-510ad0a78890","Type":"ContainerStarted","Data":"d0d032c46ac164fc69a1d0f31ab9f1395815e03227937de02e99a5eb98ddeeaa"} Nov 23 07:36:09 crc kubenswrapper[4559]: I1123 07:36:09.273799 4559 scope.go:117] "RemoveContainer" containerID="c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4" Nov 23 07:36:09 crc kubenswrapper[4559]: E1123 07:36:09.274844 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:36:10 crc kubenswrapper[4559]: I1123 07:36:10.164922 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jflsv" Nov 23 07:36:10 crc kubenswrapper[4559]: I1123 07:36:10.166269 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jflsv" Nov 23 07:36:10 crc kubenswrapper[4559]: I1123 07:36:10.198188 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jflsv" Nov 23 07:36:10 crc kubenswrapper[4559]: I1123 07:36:10.215546 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jflsv" podStartSLOduration=8.728808268 podStartE2EDuration="11.215524651s" podCreationTimestamp="2025-11-23 07:35:59 +0000 UTC" firstStartedPulling="2025-11-23 07:36:01.588470807 +0000 UTC m=+3063.610456422" lastFinishedPulling="2025-11-23 07:36:04.0751872 +0000 UTC m=+3066.097172805" observedRunningTime="2025-11-23 07:36:04.630031712 +0000 UTC m=+3066.652017326" watchObservedRunningTime="2025-11-23 07:36:10.215524651 +0000 UTC m=+3072.237510266" Nov 23 07:36:10 crc kubenswrapper[4559]: I1123 07:36:10.683699 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jflsv" Nov 23 07:36:10 crc kubenswrapper[4559]: I1123 07:36:10.717255 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jflsv"] Nov 23 07:36:12 crc kubenswrapper[4559]: I1123 07:36:12.664941 4559 generic.go:334] "Generic (PLEG): container finished" podID="51c39ef5-3bee-4998-b67c-f9f51aa5ec48" containerID="c2f7210eb3f838b99150da4d583427f03f082a736f9a65fbfe42d31a2659d3a9" exitCode=0 Nov 23 07:36:12 crc kubenswrapper[4559]: I1123 07:36:12.665291 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jflsv" podUID="0cb21f8e-86bb-48f1-9510-510ad0a78890" containerName="registry-server" containerID="cri-o://d0d032c46ac164fc69a1d0f31ab9f1395815e03227937de02e99a5eb98ddeeaa" gracePeriod=2 Nov 23 07:36:12 crc kubenswrapper[4559]: I1123 07:36:12.665117 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m5msz/crc-debug-426z9" event={"ID":"51c39ef5-3bee-4998-b67c-f9f51aa5ec48","Type":"ContainerDied","Data":"c2f7210eb3f838b99150da4d583427f03f082a736f9a65fbfe42d31a2659d3a9"} Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.013665 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jflsv" Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.196873 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cb21f8e-86bb-48f1-9510-510ad0a78890-catalog-content\") pod \"0cb21f8e-86bb-48f1-9510-510ad0a78890\" (UID: \"0cb21f8e-86bb-48f1-9510-510ad0a78890\") " Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.196928 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4zc6\" (UniqueName: \"kubernetes.io/projected/0cb21f8e-86bb-48f1-9510-510ad0a78890-kube-api-access-z4zc6\") pod \"0cb21f8e-86bb-48f1-9510-510ad0a78890\" (UID: \"0cb21f8e-86bb-48f1-9510-510ad0a78890\") " Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.196978 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cb21f8e-86bb-48f1-9510-510ad0a78890-utilities\") pod \"0cb21f8e-86bb-48f1-9510-510ad0a78890\" (UID: \"0cb21f8e-86bb-48f1-9510-510ad0a78890\") " Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.197749 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cb21f8e-86bb-48f1-9510-510ad0a78890-utilities" (OuterVolumeSpecName: "utilities") pod "0cb21f8e-86bb-48f1-9510-510ad0a78890" (UID: "0cb21f8e-86bb-48f1-9510-510ad0a78890"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.201047 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cb21f8e-86bb-48f1-9510-510ad0a78890-kube-api-access-z4zc6" (OuterVolumeSpecName: "kube-api-access-z4zc6") pod "0cb21f8e-86bb-48f1-9510-510ad0a78890" (UID: "0cb21f8e-86bb-48f1-9510-510ad0a78890"). InnerVolumeSpecName "kube-api-access-z4zc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.301255 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4zc6\" (UniqueName: \"kubernetes.io/projected/0cb21f8e-86bb-48f1-9510-510ad0a78890-kube-api-access-z4zc6\") on node \"crc\" DevicePath \"\"" Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.301287 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cb21f8e-86bb-48f1-9510-510ad0a78890-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.673796 4559 generic.go:334] "Generic (PLEG): container finished" podID="0cb21f8e-86bb-48f1-9510-510ad0a78890" containerID="d0d032c46ac164fc69a1d0f31ab9f1395815e03227937de02e99a5eb98ddeeaa" exitCode=0 Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.673863 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jflsv" Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.673866 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jflsv" event={"ID":"0cb21f8e-86bb-48f1-9510-510ad0a78890","Type":"ContainerDied","Data":"d0d032c46ac164fc69a1d0f31ab9f1395815e03227937de02e99a5eb98ddeeaa"} Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.674126 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jflsv" event={"ID":"0cb21f8e-86bb-48f1-9510-510ad0a78890","Type":"ContainerDied","Data":"cacce27b257e7f5afc70eaa69665d0c774ad78989946bb06931d228f6686755a"} Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.674161 4559 scope.go:117] "RemoveContainer" containerID="d0d032c46ac164fc69a1d0f31ab9f1395815e03227937de02e99a5eb98ddeeaa" Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.722329 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m5msz/crc-debug-426z9" Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.724034 4559 scope.go:117] "RemoveContainer" containerID="9f531680a4968a698989537855c95acb162cd99751a3d1c3ab35605d0e818ce5" Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.744606 4559 scope.go:117] "RemoveContainer" containerID="f4321a1001380e6fe72d86292078c37a1eac98137c8bdc9165dd64e3a482443b" Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.746456 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-m5msz/crc-debug-426z9"] Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.755517 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-m5msz/crc-debug-426z9"] Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.760393 4559 scope.go:117] "RemoveContainer" containerID="d0d032c46ac164fc69a1d0f31ab9f1395815e03227937de02e99a5eb98ddeeaa" Nov 23 07:36:13 crc kubenswrapper[4559]: E1123 07:36:13.761812 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0d032c46ac164fc69a1d0f31ab9f1395815e03227937de02e99a5eb98ddeeaa\": container with ID starting with d0d032c46ac164fc69a1d0f31ab9f1395815e03227937de02e99a5eb98ddeeaa not found: ID does not exist" containerID="d0d032c46ac164fc69a1d0f31ab9f1395815e03227937de02e99a5eb98ddeeaa" Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.761844 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0d032c46ac164fc69a1d0f31ab9f1395815e03227937de02e99a5eb98ddeeaa"} err="failed to get container status \"d0d032c46ac164fc69a1d0f31ab9f1395815e03227937de02e99a5eb98ddeeaa\": rpc error: code = NotFound desc = could not find container \"d0d032c46ac164fc69a1d0f31ab9f1395815e03227937de02e99a5eb98ddeeaa\": container with ID starting with d0d032c46ac164fc69a1d0f31ab9f1395815e03227937de02e99a5eb98ddeeaa not found: ID does not exist" Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.761864 4559 scope.go:117] "RemoveContainer" containerID="9f531680a4968a698989537855c95acb162cd99751a3d1c3ab35605d0e818ce5" Nov 23 07:36:13 crc kubenswrapper[4559]: E1123 07:36:13.762138 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f531680a4968a698989537855c95acb162cd99751a3d1c3ab35605d0e818ce5\": container with ID starting with 9f531680a4968a698989537855c95acb162cd99751a3d1c3ab35605d0e818ce5 not found: ID does not exist" containerID="9f531680a4968a698989537855c95acb162cd99751a3d1c3ab35605d0e818ce5" Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.762157 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f531680a4968a698989537855c95acb162cd99751a3d1c3ab35605d0e818ce5"} err="failed to get container status \"9f531680a4968a698989537855c95acb162cd99751a3d1c3ab35605d0e818ce5\": rpc error: code = NotFound desc = could not find container \"9f531680a4968a698989537855c95acb162cd99751a3d1c3ab35605d0e818ce5\": container with ID starting with 9f531680a4968a698989537855c95acb162cd99751a3d1c3ab35605d0e818ce5 not found: ID does not exist" Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.762170 4559 scope.go:117] "RemoveContainer" containerID="f4321a1001380e6fe72d86292078c37a1eac98137c8bdc9165dd64e3a482443b" Nov 23 07:36:13 crc kubenswrapper[4559]: E1123 07:36:13.762423 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4321a1001380e6fe72d86292078c37a1eac98137c8bdc9165dd64e3a482443b\": container with ID starting with f4321a1001380e6fe72d86292078c37a1eac98137c8bdc9165dd64e3a482443b not found: ID does not exist" containerID="f4321a1001380e6fe72d86292078c37a1eac98137c8bdc9165dd64e3a482443b" Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.762464 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4321a1001380e6fe72d86292078c37a1eac98137c8bdc9165dd64e3a482443b"} err="failed to get container status \"f4321a1001380e6fe72d86292078c37a1eac98137c8bdc9165dd64e3a482443b\": rpc error: code = NotFound desc = could not find container \"f4321a1001380e6fe72d86292078c37a1eac98137c8bdc9165dd64e3a482443b\": container with ID starting with f4321a1001380e6fe72d86292078c37a1eac98137c8bdc9165dd64e3a482443b not found: ID does not exist" Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.911869 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/51c39ef5-3bee-4998-b67c-f9f51aa5ec48-host\") pod \"51c39ef5-3bee-4998-b67c-f9f51aa5ec48\" (UID: \"51c39ef5-3bee-4998-b67c-f9f51aa5ec48\") " Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.911986 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/51c39ef5-3bee-4998-b67c-f9f51aa5ec48-host" (OuterVolumeSpecName: "host") pod "51c39ef5-3bee-4998-b67c-f9f51aa5ec48" (UID: "51c39ef5-3bee-4998-b67c-f9f51aa5ec48"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.912003 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5nk9\" (UniqueName: \"kubernetes.io/projected/51c39ef5-3bee-4998-b67c-f9f51aa5ec48-kube-api-access-m5nk9\") pod \"51c39ef5-3bee-4998-b67c-f9f51aa5ec48\" (UID: \"51c39ef5-3bee-4998-b67c-f9f51aa5ec48\") " Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.912661 4559 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/51c39ef5-3bee-4998-b67c-f9f51aa5ec48-host\") on node \"crc\" DevicePath \"\"" Nov 23 07:36:13 crc kubenswrapper[4559]: I1123 07:36:13.916229 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51c39ef5-3bee-4998-b67c-f9f51aa5ec48-kube-api-access-m5nk9" (OuterVolumeSpecName: "kube-api-access-m5nk9") pod "51c39ef5-3bee-4998-b67c-f9f51aa5ec48" (UID: "51c39ef5-3bee-4998-b67c-f9f51aa5ec48"). InnerVolumeSpecName "kube-api-access-m5nk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:36:14 crc kubenswrapper[4559]: I1123 07:36:14.016241 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5nk9\" (UniqueName: \"kubernetes.io/projected/51c39ef5-3bee-4998-b67c-f9f51aa5ec48-kube-api-access-m5nk9\") on node \"crc\" DevicePath \"\"" Nov 23 07:36:14 crc kubenswrapper[4559]: I1123 07:36:14.055996 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cb21f8e-86bb-48f1-9510-510ad0a78890-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0cb21f8e-86bb-48f1-9510-510ad0a78890" (UID: "0cb21f8e-86bb-48f1-9510-510ad0a78890"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:36:14 crc kubenswrapper[4559]: I1123 07:36:14.117656 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cb21f8e-86bb-48f1-9510-510ad0a78890-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:36:14 crc kubenswrapper[4559]: I1123 07:36:14.281923 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51c39ef5-3bee-4998-b67c-f9f51aa5ec48" path="/var/lib/kubelet/pods/51c39ef5-3bee-4998-b67c-f9f51aa5ec48/volumes" Nov 23 07:36:14 crc kubenswrapper[4559]: I1123 07:36:14.300622 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jflsv"] Nov 23 07:36:14 crc kubenswrapper[4559]: I1123 07:36:14.308325 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jflsv"] Nov 23 07:36:14 crc kubenswrapper[4559]: I1123 07:36:14.683237 4559 scope.go:117] "RemoveContainer" containerID="c2f7210eb3f838b99150da4d583427f03f082a736f9a65fbfe42d31a2659d3a9" Nov 23 07:36:14 crc kubenswrapper[4559]: I1123 07:36:14.683274 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m5msz/crc-debug-426z9" Nov 23 07:36:14 crc kubenswrapper[4559]: I1123 07:36:14.870227 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-m5msz/crc-debug-qnkml"] Nov 23 07:36:14 crc kubenswrapper[4559]: E1123 07:36:14.870519 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cb21f8e-86bb-48f1-9510-510ad0a78890" containerName="extract-content" Nov 23 07:36:14 crc kubenswrapper[4559]: I1123 07:36:14.870536 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cb21f8e-86bb-48f1-9510-510ad0a78890" containerName="extract-content" Nov 23 07:36:14 crc kubenswrapper[4559]: E1123 07:36:14.870552 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cb21f8e-86bb-48f1-9510-510ad0a78890" containerName="registry-server" Nov 23 07:36:14 crc kubenswrapper[4559]: I1123 07:36:14.870558 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cb21f8e-86bb-48f1-9510-510ad0a78890" containerName="registry-server" Nov 23 07:36:14 crc kubenswrapper[4559]: E1123 07:36:14.870579 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51c39ef5-3bee-4998-b67c-f9f51aa5ec48" containerName="container-00" Nov 23 07:36:14 crc kubenswrapper[4559]: I1123 07:36:14.870586 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="51c39ef5-3bee-4998-b67c-f9f51aa5ec48" containerName="container-00" Nov 23 07:36:14 crc kubenswrapper[4559]: E1123 07:36:14.870596 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cb21f8e-86bb-48f1-9510-510ad0a78890" containerName="extract-utilities" Nov 23 07:36:14 crc kubenswrapper[4559]: I1123 07:36:14.870601 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cb21f8e-86bb-48f1-9510-510ad0a78890" containerName="extract-utilities" Nov 23 07:36:14 crc kubenswrapper[4559]: I1123 07:36:14.870772 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="51c39ef5-3bee-4998-b67c-f9f51aa5ec48" containerName="container-00" Nov 23 07:36:14 crc kubenswrapper[4559]: I1123 07:36:14.870794 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cb21f8e-86bb-48f1-9510-510ad0a78890" containerName="registry-server" Nov 23 07:36:14 crc kubenswrapper[4559]: I1123 07:36:14.871286 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m5msz/crc-debug-qnkml" Nov 23 07:36:14 crc kubenswrapper[4559]: I1123 07:36:14.927677 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97kqm\" (UniqueName: \"kubernetes.io/projected/cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa-kube-api-access-97kqm\") pod \"crc-debug-qnkml\" (UID: \"cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa\") " pod="openshift-must-gather-m5msz/crc-debug-qnkml" Nov 23 07:36:14 crc kubenswrapper[4559]: I1123 07:36:14.927728 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa-host\") pod \"crc-debug-qnkml\" (UID: \"cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa\") " pod="openshift-must-gather-m5msz/crc-debug-qnkml" Nov 23 07:36:15 crc kubenswrapper[4559]: I1123 07:36:15.029804 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97kqm\" (UniqueName: \"kubernetes.io/projected/cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa-kube-api-access-97kqm\") pod \"crc-debug-qnkml\" (UID: \"cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa\") " pod="openshift-must-gather-m5msz/crc-debug-qnkml" Nov 23 07:36:15 crc kubenswrapper[4559]: I1123 07:36:15.029885 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa-host\") pod \"crc-debug-qnkml\" (UID: \"cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa\") " pod="openshift-must-gather-m5msz/crc-debug-qnkml" Nov 23 07:36:15 crc kubenswrapper[4559]: I1123 07:36:15.029970 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa-host\") pod \"crc-debug-qnkml\" (UID: \"cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa\") " pod="openshift-must-gather-m5msz/crc-debug-qnkml" Nov 23 07:36:15 crc kubenswrapper[4559]: I1123 07:36:15.044337 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97kqm\" (UniqueName: \"kubernetes.io/projected/cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa-kube-api-access-97kqm\") pod \"crc-debug-qnkml\" (UID: \"cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa\") " pod="openshift-must-gather-m5msz/crc-debug-qnkml" Nov 23 07:36:15 crc kubenswrapper[4559]: I1123 07:36:15.183969 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m5msz/crc-debug-qnkml" Nov 23 07:36:15 crc kubenswrapper[4559]: I1123 07:36:15.692862 4559 generic.go:334] "Generic (PLEG): container finished" podID="cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa" containerID="ea6067e9c428854febcc0d77fbe1dc44e89550721a45263d660dcdc34bab7ce7" exitCode=0 Nov 23 07:36:15 crc kubenswrapper[4559]: I1123 07:36:15.692938 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m5msz/crc-debug-qnkml" event={"ID":"cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa","Type":"ContainerDied","Data":"ea6067e9c428854febcc0d77fbe1dc44e89550721a45263d660dcdc34bab7ce7"} Nov 23 07:36:15 crc kubenswrapper[4559]: I1123 07:36:15.693115 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m5msz/crc-debug-qnkml" event={"ID":"cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa","Type":"ContainerStarted","Data":"d9eba96c31f885887732eb82bedbe43aec97d1a4491469545a9eb4bf7ff510a8"} Nov 23 07:36:16 crc kubenswrapper[4559]: I1123 07:36:16.011136 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-m5msz/crc-debug-qnkml"] Nov 23 07:36:16 crc kubenswrapper[4559]: I1123 07:36:16.016871 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-m5msz/crc-debug-qnkml"] Nov 23 07:36:16 crc kubenswrapper[4559]: I1123 07:36:16.280950 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cb21f8e-86bb-48f1-9510-510ad0a78890" path="/var/lib/kubelet/pods/0cb21f8e-86bb-48f1-9510-510ad0a78890/volumes" Nov 23 07:36:16 crc kubenswrapper[4559]: I1123 07:36:16.784312 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m5msz/crc-debug-qnkml" Nov 23 07:36:16 crc kubenswrapper[4559]: I1123 07:36:16.855783 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97kqm\" (UniqueName: \"kubernetes.io/projected/cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa-kube-api-access-97kqm\") pod \"cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa\" (UID: \"cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa\") " Nov 23 07:36:16 crc kubenswrapper[4559]: I1123 07:36:16.855868 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa-host\") pod \"cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa\" (UID: \"cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa\") " Nov 23 07:36:16 crc kubenswrapper[4559]: I1123 07:36:16.856053 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa-host" (OuterVolumeSpecName: "host") pod "cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa" (UID: "cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:36:16 crc kubenswrapper[4559]: I1123 07:36:16.856283 4559 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa-host\") on node \"crc\" DevicePath \"\"" Nov 23 07:36:16 crc kubenswrapper[4559]: I1123 07:36:16.860143 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa-kube-api-access-97kqm" (OuterVolumeSpecName: "kube-api-access-97kqm") pod "cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa" (UID: "cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa"). InnerVolumeSpecName "kube-api-access-97kqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:36:16 crc kubenswrapper[4559]: I1123 07:36:16.958362 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97kqm\" (UniqueName: \"kubernetes.io/projected/cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa-kube-api-access-97kqm\") on node \"crc\" DevicePath \"\"" Nov 23 07:36:17 crc kubenswrapper[4559]: I1123 07:36:17.130403 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-m5msz/crc-debug-9b94s"] Nov 23 07:36:17 crc kubenswrapper[4559]: E1123 07:36:17.130747 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa" containerName="container-00" Nov 23 07:36:17 crc kubenswrapper[4559]: I1123 07:36:17.130766 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa" containerName="container-00" Nov 23 07:36:17 crc kubenswrapper[4559]: I1123 07:36:17.130950 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa" containerName="container-00" Nov 23 07:36:17 crc kubenswrapper[4559]: I1123 07:36:17.131506 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m5msz/crc-debug-9b94s" Nov 23 07:36:17 crc kubenswrapper[4559]: I1123 07:36:17.162358 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnc7k\" (UniqueName: \"kubernetes.io/projected/a9b6f723-e379-4cfb-8327-ae8289d9cf73-kube-api-access-pnc7k\") pod \"crc-debug-9b94s\" (UID: \"a9b6f723-e379-4cfb-8327-ae8289d9cf73\") " pod="openshift-must-gather-m5msz/crc-debug-9b94s" Nov 23 07:36:17 crc kubenswrapper[4559]: I1123 07:36:17.162530 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a9b6f723-e379-4cfb-8327-ae8289d9cf73-host\") pod \"crc-debug-9b94s\" (UID: \"a9b6f723-e379-4cfb-8327-ae8289d9cf73\") " pod="openshift-must-gather-m5msz/crc-debug-9b94s" Nov 23 07:36:17 crc kubenswrapper[4559]: I1123 07:36:17.263809 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a9b6f723-e379-4cfb-8327-ae8289d9cf73-host\") pod \"crc-debug-9b94s\" (UID: \"a9b6f723-e379-4cfb-8327-ae8289d9cf73\") " pod="openshift-must-gather-m5msz/crc-debug-9b94s" Nov 23 07:36:17 crc kubenswrapper[4559]: I1123 07:36:17.263962 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a9b6f723-e379-4cfb-8327-ae8289d9cf73-host\") pod \"crc-debug-9b94s\" (UID: \"a9b6f723-e379-4cfb-8327-ae8289d9cf73\") " pod="openshift-must-gather-m5msz/crc-debug-9b94s" Nov 23 07:36:17 crc kubenswrapper[4559]: I1123 07:36:17.264108 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnc7k\" (UniqueName: \"kubernetes.io/projected/a9b6f723-e379-4cfb-8327-ae8289d9cf73-kube-api-access-pnc7k\") pod \"crc-debug-9b94s\" (UID: \"a9b6f723-e379-4cfb-8327-ae8289d9cf73\") " pod="openshift-must-gather-m5msz/crc-debug-9b94s" Nov 23 07:36:17 crc kubenswrapper[4559]: I1123 07:36:17.277728 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnc7k\" (UniqueName: \"kubernetes.io/projected/a9b6f723-e379-4cfb-8327-ae8289d9cf73-kube-api-access-pnc7k\") pod \"crc-debug-9b94s\" (UID: \"a9b6f723-e379-4cfb-8327-ae8289d9cf73\") " pod="openshift-must-gather-m5msz/crc-debug-9b94s" Nov 23 07:36:17 crc kubenswrapper[4559]: I1123 07:36:17.444978 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m5msz/crc-debug-9b94s" Nov 23 07:36:17 crc kubenswrapper[4559]: W1123 07:36:17.476382 4559 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9b6f723_e379_4cfb_8327_ae8289d9cf73.slice/crio-8ed9182f8c791eea73d58839a1f8db639b0fbc5d1d1ace82c9ac1f1f8732d893 WatchSource:0}: Error finding container 8ed9182f8c791eea73d58839a1f8db639b0fbc5d1d1ace82c9ac1f1f8732d893: Status 404 returned error can't find the container with id 8ed9182f8c791eea73d58839a1f8db639b0fbc5d1d1ace82c9ac1f1f8732d893 Nov 23 07:36:17 crc kubenswrapper[4559]: I1123 07:36:17.707027 4559 scope.go:117] "RemoveContainer" containerID="ea6067e9c428854febcc0d77fbe1dc44e89550721a45263d660dcdc34bab7ce7" Nov 23 07:36:17 crc kubenswrapper[4559]: I1123 07:36:17.707030 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m5msz/crc-debug-qnkml" Nov 23 07:36:17 crc kubenswrapper[4559]: I1123 07:36:17.708879 4559 generic.go:334] "Generic (PLEG): container finished" podID="a9b6f723-e379-4cfb-8327-ae8289d9cf73" containerID="ccb3fc015b2c0202c04d80333ba56aacc3138faa92d0a28d2a07f3cc1bbafd9f" exitCode=0 Nov 23 07:36:17 crc kubenswrapper[4559]: I1123 07:36:17.708927 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m5msz/crc-debug-9b94s" event={"ID":"a9b6f723-e379-4cfb-8327-ae8289d9cf73","Type":"ContainerDied","Data":"ccb3fc015b2c0202c04d80333ba56aacc3138faa92d0a28d2a07f3cc1bbafd9f"} Nov 23 07:36:17 crc kubenswrapper[4559]: I1123 07:36:17.708949 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m5msz/crc-debug-9b94s" event={"ID":"a9b6f723-e379-4cfb-8327-ae8289d9cf73","Type":"ContainerStarted","Data":"8ed9182f8c791eea73d58839a1f8db639b0fbc5d1d1ace82c9ac1f1f8732d893"} Nov 23 07:36:17 crc kubenswrapper[4559]: I1123 07:36:17.737636 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-m5msz/crc-debug-9b94s"] Nov 23 07:36:17 crc kubenswrapper[4559]: I1123 07:36:17.745460 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-m5msz/crc-debug-9b94s"] Nov 23 07:36:18 crc kubenswrapper[4559]: I1123 07:36:18.281734 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa" path="/var/lib/kubelet/pods/cb6e3c1b-7dea-4a4a-8227-3722e0b8d6aa/volumes" Nov 23 07:36:18 crc kubenswrapper[4559]: I1123 07:36:18.786899 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m5msz/crc-debug-9b94s" Nov 23 07:36:18 crc kubenswrapper[4559]: I1123 07:36:18.888694 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a9b6f723-e379-4cfb-8327-ae8289d9cf73-host\") pod \"a9b6f723-e379-4cfb-8327-ae8289d9cf73\" (UID: \"a9b6f723-e379-4cfb-8327-ae8289d9cf73\") " Nov 23 07:36:18 crc kubenswrapper[4559]: I1123 07:36:18.888817 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9b6f723-e379-4cfb-8327-ae8289d9cf73-host" (OuterVolumeSpecName: "host") pod "a9b6f723-e379-4cfb-8327-ae8289d9cf73" (UID: "a9b6f723-e379-4cfb-8327-ae8289d9cf73"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 07:36:18 crc kubenswrapper[4559]: I1123 07:36:18.888835 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnc7k\" (UniqueName: \"kubernetes.io/projected/a9b6f723-e379-4cfb-8327-ae8289d9cf73-kube-api-access-pnc7k\") pod \"a9b6f723-e379-4cfb-8327-ae8289d9cf73\" (UID: \"a9b6f723-e379-4cfb-8327-ae8289d9cf73\") " Nov 23 07:36:18 crc kubenswrapper[4559]: I1123 07:36:18.889235 4559 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a9b6f723-e379-4cfb-8327-ae8289d9cf73-host\") on node \"crc\" DevicePath \"\"" Nov 23 07:36:18 crc kubenswrapper[4559]: I1123 07:36:18.893211 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9b6f723-e379-4cfb-8327-ae8289d9cf73-kube-api-access-pnc7k" (OuterVolumeSpecName: "kube-api-access-pnc7k") pod "a9b6f723-e379-4cfb-8327-ae8289d9cf73" (UID: "a9b6f723-e379-4cfb-8327-ae8289d9cf73"). InnerVolumeSpecName "kube-api-access-pnc7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:36:18 crc kubenswrapper[4559]: I1123 07:36:18.990471 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnc7k\" (UniqueName: \"kubernetes.io/projected/a9b6f723-e379-4cfb-8327-ae8289d9cf73-kube-api-access-pnc7k\") on node \"crc\" DevicePath \"\"" Nov 23 07:36:19 crc kubenswrapper[4559]: I1123 07:36:19.724774 4559 scope.go:117] "RemoveContainer" containerID="ccb3fc015b2c0202c04d80333ba56aacc3138faa92d0a28d2a07f3cc1bbafd9f" Nov 23 07:36:19 crc kubenswrapper[4559]: I1123 07:36:19.724811 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m5msz/crc-debug-9b94s" Nov 23 07:36:20 crc kubenswrapper[4559]: I1123 07:36:20.281582 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9b6f723-e379-4cfb-8327-ae8289d9cf73" path="/var/lib/kubelet/pods/a9b6f723-e379-4cfb-8327-ae8289d9cf73/volumes" Nov 23 07:36:24 crc kubenswrapper[4559]: I1123 07:36:24.273592 4559 scope.go:117] "RemoveContainer" containerID="c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4" Nov 23 07:36:24 crc kubenswrapper[4559]: E1123 07:36:24.274192 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:36:32 crc kubenswrapper[4559]: I1123 07:36:32.979308 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-55bb64d698-6lpkd_f9264f09-464d-404a-88a1-7f86ecb2ccb8/barbican-api/0.log" Nov 23 07:36:33 crc kubenswrapper[4559]: I1123 07:36:33.074050 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-55bb64d698-6lpkd_f9264f09-464d-404a-88a1-7f86ecb2ccb8/barbican-api-log/0.log" Nov 23 07:36:33 crc kubenswrapper[4559]: I1123 07:36:33.161395 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-685df6659d-zwp5c_cb7ba735-691f-4eb5-ac0c-c3f976e8e604/barbican-keystone-listener-log/0.log" Nov 23 07:36:33 crc kubenswrapper[4559]: I1123 07:36:33.206129 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-685df6659d-zwp5c_cb7ba735-691f-4eb5-ac0c-c3f976e8e604/barbican-keystone-listener/0.log" Nov 23 07:36:33 crc kubenswrapper[4559]: I1123 07:36:33.312677 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-d4dd9b9cc-ndjpd_f5b2ad32-159b-4369-951d-02d3ae8581eb/barbican-worker/0.log" Nov 23 07:36:33 crc kubenswrapper[4559]: I1123 07:36:33.331167 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-d4dd9b9cc-ndjpd_f5b2ad32-159b-4369-951d-02d3ae8581eb/barbican-worker-log/0.log" Nov 23 07:36:33 crc kubenswrapper[4559]: I1123 07:36:33.426965 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-ttlkz_af59bdf5-b08e-487d-bcc4-c70a275cace5/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:36:33 crc kubenswrapper[4559]: I1123 07:36:33.514424 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_865e8bed-9dfe-4f61-9c40-e2876af8f95b/ceilometer-central-agent/0.log" Nov 23 07:36:33 crc kubenswrapper[4559]: I1123 07:36:33.575551 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_865e8bed-9dfe-4f61-9c40-e2876af8f95b/ceilometer-notification-agent/0.log" Nov 23 07:36:33 crc kubenswrapper[4559]: I1123 07:36:33.589014 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_865e8bed-9dfe-4f61-9c40-e2876af8f95b/proxy-httpd/0.log" Nov 23 07:36:33 crc kubenswrapper[4559]: I1123 07:36:33.655445 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_865e8bed-9dfe-4f61-9c40-e2876af8f95b/sg-core/0.log" Nov 23 07:36:33 crc kubenswrapper[4559]: I1123 07:36:33.736522 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_3b8ac924-b4b7-4374-8f5f-7b2133936385/cinder-api-log/0.log" Nov 23 07:36:33 crc kubenswrapper[4559]: I1123 07:36:33.757084 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_3b8ac924-b4b7-4374-8f5f-7b2133936385/cinder-api/0.log" Nov 23 07:36:33 crc kubenswrapper[4559]: I1123 07:36:33.894539 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_cd7f1662-fa90-4677-9c1a-e5882f079497/cinder-scheduler/0.log" Nov 23 07:36:33 crc kubenswrapper[4559]: I1123 07:36:33.928149 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_cd7f1662-fa90-4677-9c1a-e5882f079497/probe/0.log" Nov 23 07:36:34 crc kubenswrapper[4559]: I1123 07:36:34.052436 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-nxmv5_45c516df-85dd-4a43-94ef-ea1381622c06/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:36:34 crc kubenswrapper[4559]: I1123 07:36:34.098806 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-pf6ms_f3cbccfd-3b9d-4d8d-aec1-702b2e494667/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:36:34 crc kubenswrapper[4559]: I1123 07:36:34.244792 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-869db7cd47-ghbfq_9b236045-a74d-4a5e-a8e3-c70530d9e6a2/init/0.log" Nov 23 07:36:34 crc kubenswrapper[4559]: I1123 07:36:34.360248 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-869db7cd47-ghbfq_9b236045-a74d-4a5e-a8e3-c70530d9e6a2/init/0.log" Nov 23 07:36:34 crc kubenswrapper[4559]: I1123 07:36:34.402582 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-869db7cd47-ghbfq_9b236045-a74d-4a5e-a8e3-c70530d9e6a2/dnsmasq-dns/0.log" Nov 23 07:36:34 crc kubenswrapper[4559]: I1123 07:36:34.458595 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-46r2s_fd4d9a1f-5591-47b2-8fb2-0d87b87e3c29/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:36:34 crc kubenswrapper[4559]: I1123 07:36:34.543152 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_24caf856-63a1-44eb-a0f2-7afc985ff668/glance-httpd/0.log" Nov 23 07:36:34 crc kubenswrapper[4559]: I1123 07:36:34.625793 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_24caf856-63a1-44eb-a0f2-7afc985ff668/glance-log/0.log" Nov 23 07:36:34 crc kubenswrapper[4559]: I1123 07:36:34.709406 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_c869f924-be1d-4e33-8a16-5a833a6ad9d0/glance-httpd/0.log" Nov 23 07:36:34 crc kubenswrapper[4559]: I1123 07:36:34.786451 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_c869f924-be1d-4e33-8a16-5a833a6ad9d0/glance-log/0.log" Nov 23 07:36:34 crc kubenswrapper[4559]: I1123 07:36:34.800558 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-vg4bs_aa380021-cb24-43b1-bb0d-e3dbbd5f6676/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:36:34 crc kubenswrapper[4559]: I1123 07:36:34.981706 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-wgnh6_460a9b32-addb-458a-9913-4ba0433b3eb3/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:36:35 crc kubenswrapper[4559]: I1123 07:36:35.151911 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-74ff86f86c-gcbcl_dca8804f-80bf-4b44-99d0-7c2a9fea9166/keystone-api/0.log" Nov 23 07:36:35 crc kubenswrapper[4559]: I1123 07:36:35.200839 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29398021-r88wd_da23948d-f4d4-42cc-9960-aaeb3efc2309/keystone-cron/0.log" Nov 23 07:36:35 crc kubenswrapper[4559]: I1123 07:36:35.274082 4559 scope.go:117] "RemoveContainer" containerID="c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4" Nov 23 07:36:35 crc kubenswrapper[4559]: E1123 07:36:35.274341 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:36:35 crc kubenswrapper[4559]: I1123 07:36:35.294858 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_916e4424-5f09-44c7-8b07-de2a4d84df18/kube-state-metrics/0.log" Nov 23 07:36:35 crc kubenswrapper[4559]: I1123 07:36:35.371222 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-njqpj_aa33c87f-c563-4944-883e-3ed5649b96a5/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:36:35 crc kubenswrapper[4559]: I1123 07:36:35.606851 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-d8c98f585-kr7vt_72027df5-f460-436a-b883-1895caea6f90/neutron-httpd/0.log" Nov 23 07:36:35 crc kubenswrapper[4559]: I1123 07:36:35.688217 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-d8c98f585-kr7vt_72027df5-f460-436a-b883-1895caea6f90/neutron-api/0.log" Nov 23 07:36:35 crc kubenswrapper[4559]: I1123 07:36:35.781996 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-l5g2n_4f9b53ba-da26-40bb-9819-cdeb54deaef2/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:36:36 crc kubenswrapper[4559]: I1123 07:36:36.244902 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e55ca2b9-735d-434f-899d-cef9bce42b2e/nova-api-log/0.log" Nov 23 07:36:36 crc kubenswrapper[4559]: I1123 07:36:36.290489 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_a5666b61-0dcd-4fcd-a96a-44fa29cd3d84/nova-cell0-conductor-conductor/0.log" Nov 23 07:36:36 crc kubenswrapper[4559]: I1123 07:36:36.499883 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e55ca2b9-735d-434f-899d-cef9bce42b2e/nova-api-api/0.log" Nov 23 07:36:36 crc kubenswrapper[4559]: I1123 07:36:36.518783 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_56c491cd-4182-4477-a9de-d7e91d68e07f/nova-cell1-conductor-conductor/0.log" Nov 23 07:36:36 crc kubenswrapper[4559]: I1123 07:36:36.578801 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_75fcd1aa-df9b-4cc4-91aa-dde8e919ff9b/nova-cell1-novncproxy-novncproxy/0.log" Nov 23 07:36:36 crc kubenswrapper[4559]: I1123 07:36:36.754861 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-dfw5h_2505b564-8639-464b-b223-91af7ab3661d/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:36:36 crc kubenswrapper[4559]: I1123 07:36:36.823466 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_66d6e144-f771-40b4-a6f2-0689654cf797/nova-metadata-log/0.log" Nov 23 07:36:37 crc kubenswrapper[4559]: I1123 07:36:37.097103 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_ba80f653-69d6-4a80-aaab-2747e0f9b61b/nova-scheduler-scheduler/0.log" Nov 23 07:36:37 crc kubenswrapper[4559]: I1123 07:36:37.153102 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_01dff7d1-bf4c-4d88-900c-99de62949ac7/mysql-bootstrap/0.log" Nov 23 07:36:37 crc kubenswrapper[4559]: I1123 07:36:37.343665 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_01dff7d1-bf4c-4d88-900c-99de62949ac7/galera/0.log" Nov 23 07:36:37 crc kubenswrapper[4559]: I1123 07:36:37.365311 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_01dff7d1-bf4c-4d88-900c-99de62949ac7/mysql-bootstrap/0.log" Nov 23 07:36:37 crc kubenswrapper[4559]: I1123 07:36:37.503955 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e097ab60-4653-4038-b6de-40b55936565c/mysql-bootstrap/0.log" Nov 23 07:36:37 crc kubenswrapper[4559]: I1123 07:36:37.664538 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e097ab60-4653-4038-b6de-40b55936565c/mysql-bootstrap/0.log" Nov 23 07:36:37 crc kubenswrapper[4559]: I1123 07:36:37.670886 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e097ab60-4653-4038-b6de-40b55936565c/galera/0.log" Nov 23 07:36:37 crc kubenswrapper[4559]: I1123 07:36:37.686349 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_66d6e144-f771-40b4-a6f2-0689654cf797/nova-metadata-metadata/0.log" Nov 23 07:36:37 crc kubenswrapper[4559]: I1123 07:36:37.789443 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_0b70547f-d8a6-43c4-b2fa-20f8d3f02bfa/openstackclient/0.log" Nov 23 07:36:37 crc kubenswrapper[4559]: I1123 07:36:37.870285 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-blj85_20662e83-ba79-4c97-80fb-98fda28c1149/ovn-controller/0.log" Nov 23 07:36:38 crc kubenswrapper[4559]: I1123 07:36:38.072850 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-nzw6k_dcfa74ae-9b9d-447c-a401-a93ae8dfa8f5/openstack-network-exporter/0.log" Nov 23 07:36:38 crc kubenswrapper[4559]: I1123 07:36:38.171794 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jg7hk_d534f227-972b-4a85-aeb5-8f1d226352e1/ovsdb-server-init/0.log" Nov 23 07:36:38 crc kubenswrapper[4559]: I1123 07:36:38.320446 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jg7hk_d534f227-972b-4a85-aeb5-8f1d226352e1/ovsdb-server-init/0.log" Nov 23 07:36:38 crc kubenswrapper[4559]: I1123 07:36:38.352218 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jg7hk_d534f227-972b-4a85-aeb5-8f1d226352e1/ovs-vswitchd/0.log" Nov 23 07:36:38 crc kubenswrapper[4559]: I1123 07:36:38.366565 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jg7hk_d534f227-972b-4a85-aeb5-8f1d226352e1/ovsdb-server/0.log" Nov 23 07:36:38 crc kubenswrapper[4559]: I1123 07:36:38.504013 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-9svlh_4f4ac16e-c282-4fa3-bdbe-e5f848782f47/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:36:38 crc kubenswrapper[4559]: I1123 07:36:38.512604 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_00fa46d4-5d8e-43a6-a182-34faf0d694ab/ovn-northd/0.log" Nov 23 07:36:38 crc kubenswrapper[4559]: I1123 07:36:38.514562 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_00fa46d4-5d8e-43a6-a182-34faf0d694ab/openstack-network-exporter/0.log" Nov 23 07:36:38 crc kubenswrapper[4559]: I1123 07:36:38.656132 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_5d19fe8b-239a-476a-9213-1eccbd58958e/openstack-network-exporter/0.log" Nov 23 07:36:38 crc kubenswrapper[4559]: I1123 07:36:38.730501 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_5d19fe8b-239a-476a-9213-1eccbd58958e/ovsdbserver-nb/0.log" Nov 23 07:36:38 crc kubenswrapper[4559]: I1123 07:36:38.819454 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f162538d-e942-4d5e-b2c5-11ba57054cf4/openstack-network-exporter/0.log" Nov 23 07:36:38 crc kubenswrapper[4559]: I1123 07:36:38.875058 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f162538d-e942-4d5e-b2c5-11ba57054cf4/ovsdbserver-sb/0.log" Nov 23 07:36:38 crc kubenswrapper[4559]: I1123 07:36:38.960591 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7667f44776-mszxt_1df68d96-dfae-42c9-8a11-9906486ea1b9/placement-api/0.log" Nov 23 07:36:39 crc kubenswrapper[4559]: I1123 07:36:39.029652 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7667f44776-mszxt_1df68d96-dfae-42c9-8a11-9906486ea1b9/placement-log/0.log" Nov 23 07:36:39 crc kubenswrapper[4559]: I1123 07:36:39.131217 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_e0731be9-7300-45df-b028-8aeb2a3999b4/setup-container/0.log" Nov 23 07:36:39 crc kubenswrapper[4559]: I1123 07:36:39.255818 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_e0731be9-7300-45df-b028-8aeb2a3999b4/setup-container/0.log" Nov 23 07:36:39 crc kubenswrapper[4559]: I1123 07:36:39.298627 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_e0731be9-7300-45df-b028-8aeb2a3999b4/rabbitmq/0.log" Nov 23 07:36:39 crc kubenswrapper[4559]: I1123 07:36:39.337074 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8/setup-container/0.log" Nov 23 07:36:39 crc kubenswrapper[4559]: I1123 07:36:39.521195 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-z4nn2_67ea0e44-543c-4601-88e5-51ee0d43424a/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:36:39 crc kubenswrapper[4559]: I1123 07:36:39.527532 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8/rabbitmq/0.log" Nov 23 07:36:39 crc kubenswrapper[4559]: I1123 07:36:39.549546 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_bb28b30b-cac7-4c9b-b02b-4c62e96b1ae8/setup-container/0.log" Nov 23 07:36:39 crc kubenswrapper[4559]: I1123 07:36:39.664048 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-sx224_1b84c748-5838-4938-be55-0e8fc4174730/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:36:39 crc kubenswrapper[4559]: I1123 07:36:39.705127 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-ntsdb_c5fa1816-8789-4600-931c-8a376f0936c0/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:36:39 crc kubenswrapper[4559]: I1123 07:36:39.867975 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-sw6gw_3f1151cd-5698-4fd5-ba25-e28b48717cb8/ssh-known-hosts-edpm-deployment/0.log" Nov 23 07:36:39 crc kubenswrapper[4559]: I1123 07:36:39.904745 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-5flwd_f91e51c1-c500-4da2-acba-fcdfa134d397/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:36:40 crc kubenswrapper[4559]: I1123 07:36:40.095009 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6b4fdd88f9-9hnng_db083799-cb94-42d3-8a54-8446f7a76502/proxy-httpd/0.log" Nov 23 07:36:40 crc kubenswrapper[4559]: I1123 07:36:40.107803 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6b4fdd88f9-9hnng_db083799-cb94-42d3-8a54-8446f7a76502/proxy-server/0.log" Nov 23 07:36:40 crc kubenswrapper[4559]: I1123 07:36:40.200655 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-94gfd_3696cad8-c9be-4efd-982a-a4c1d6de858b/swift-ring-rebalance/0.log" Nov 23 07:36:40 crc kubenswrapper[4559]: I1123 07:36:40.306284 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/account-auditor/0.log" Nov 23 07:36:40 crc kubenswrapper[4559]: I1123 07:36:40.317763 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/account-reaper/0.log" Nov 23 07:36:40 crc kubenswrapper[4559]: I1123 07:36:40.399449 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/account-replicator/0.log" Nov 23 07:36:40 crc kubenswrapper[4559]: I1123 07:36:40.459432 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/account-server/0.log" Nov 23 07:36:40 crc kubenswrapper[4559]: I1123 07:36:40.482536 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/container-auditor/0.log" Nov 23 07:36:40 crc kubenswrapper[4559]: I1123 07:36:40.513821 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/container-replicator/0.log" Nov 23 07:36:40 crc kubenswrapper[4559]: I1123 07:36:40.575195 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/container-server/0.log" Nov 23 07:36:40 crc kubenswrapper[4559]: I1123 07:36:40.640487 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/object-auditor/0.log" Nov 23 07:36:40 crc kubenswrapper[4559]: I1123 07:36:40.656198 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/container-updater/0.log" Nov 23 07:36:40 crc kubenswrapper[4559]: I1123 07:36:40.698691 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/object-expirer/0.log" Nov 23 07:36:40 crc kubenswrapper[4559]: I1123 07:36:40.760097 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/object-replicator/0.log" Nov 23 07:36:40 crc kubenswrapper[4559]: I1123 07:36:40.796540 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/object-server/0.log" Nov 23 07:36:40 crc kubenswrapper[4559]: I1123 07:36:40.833044 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/object-updater/0.log" Nov 23 07:36:40 crc kubenswrapper[4559]: I1123 07:36:40.845450 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/rsync/0.log" Nov 23 07:36:40 crc kubenswrapper[4559]: I1123 07:36:40.942417 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_7783a8fe-c77b-4e05-8c52-3e11d3a92196/swift-recon-cron/0.log" Nov 23 07:36:41 crc kubenswrapper[4559]: I1123 07:36:41.073685 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-4f7m8_14f31bb2-9f00-4931-a84a-900401189cc0/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:36:41 crc kubenswrapper[4559]: I1123 07:36:41.112768 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_fb299f6e-9a30-4d29-95db-a0cbceb573ea/tempest-tests-tempest-tests-runner/0.log" Nov 23 07:36:41 crc kubenswrapper[4559]: I1123 07:36:41.210271 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_83a546c2-d73a-40cd-aae5-22eb4aef5268/test-operator-logs-container/0.log" Nov 23 07:36:41 crc kubenswrapper[4559]: I1123 07:36:41.325823 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-crmzz_2368491a-9b45-4c74-bb8d-b6be5fca89f8/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 07:36:47 crc kubenswrapper[4559]: I1123 07:36:47.273157 4559 scope.go:117] "RemoveContainer" containerID="c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4" Nov 23 07:36:47 crc kubenswrapper[4559]: E1123 07:36:47.274025 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:36:48 crc kubenswrapper[4559]: I1123 07:36:48.849661 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_de0cb0c4-ffde-44c5-adb9-7aea0692c1b5/memcached/0.log" Nov 23 07:36:58 crc kubenswrapper[4559]: I1123 07:36:58.477733 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt_aaa08108-3039-4f4c-b01f-93dc7deefaa5/util/0.log" Nov 23 07:36:58 crc kubenswrapper[4559]: I1123 07:36:58.598038 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt_aaa08108-3039-4f4c-b01f-93dc7deefaa5/pull/0.log" Nov 23 07:36:58 crc kubenswrapper[4559]: I1123 07:36:58.601414 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt_aaa08108-3039-4f4c-b01f-93dc7deefaa5/util/0.log" Nov 23 07:36:58 crc kubenswrapper[4559]: I1123 07:36:58.623945 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt_aaa08108-3039-4f4c-b01f-93dc7deefaa5/pull/0.log" Nov 23 07:36:58 crc kubenswrapper[4559]: I1123 07:36:58.760506 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt_aaa08108-3039-4f4c-b01f-93dc7deefaa5/pull/0.log" Nov 23 07:36:58 crc kubenswrapper[4559]: I1123 07:36:58.771833 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt_aaa08108-3039-4f4c-b01f-93dc7deefaa5/util/0.log" Nov 23 07:36:58 crc kubenswrapper[4559]: I1123 07:36:58.778928 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287gbbbt_aaa08108-3039-4f4c-b01f-93dc7deefaa5/extract/0.log" Nov 23 07:36:58 crc kubenswrapper[4559]: I1123 07:36:58.906796 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7768f8c84f-ks2b7_090f1d86-b4e0-43ca-9aee-49c771cab8c1/kube-rbac-proxy/0.log" Nov 23 07:36:58 crc kubenswrapper[4559]: I1123 07:36:58.917171 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6d8fd67bf7-d78br_35944557-7948-4d54-8225-788d51eb01d1/kube-rbac-proxy/0.log" Nov 23 07:36:58 crc kubenswrapper[4559]: I1123 07:36:58.975121 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7768f8c84f-ks2b7_090f1d86-b4e0-43ca-9aee-49c771cab8c1/manager/0.log" Nov 23 07:36:59 crc kubenswrapper[4559]: I1123 07:36:59.074400 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6d8fd67bf7-d78br_35944557-7948-4d54-8225-788d51eb01d1/manager/0.log" Nov 23 07:36:59 crc kubenswrapper[4559]: I1123 07:36:59.114042 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-56dfb6b67f-47bnx_6f79eaa5-12ee-4c47-8c5f-c0fa9190bc2d/kube-rbac-proxy/0.log" Nov 23 07:36:59 crc kubenswrapper[4559]: I1123 07:36:59.161662 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-56dfb6b67f-47bnx_6f79eaa5-12ee-4c47-8c5f-c0fa9190bc2d/manager/0.log" Nov 23 07:36:59 crc kubenswrapper[4559]: I1123 07:36:59.266570 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8667fbf6f6-qfgkh_042a59c3-a00e-4934-b16e-c731f0ab903f/kube-rbac-proxy/0.log" Nov 23 07:36:59 crc kubenswrapper[4559]: I1123 07:36:59.313655 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8667fbf6f6-qfgkh_042a59c3-a00e-4934-b16e-c731f0ab903f/manager/0.log" Nov 23 07:36:59 crc kubenswrapper[4559]: I1123 07:36:59.424462 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-bf4c6585d-9tvg2_5b70e018-1ab3-4f23-b43f-1f433a096b81/kube-rbac-proxy/0.log" Nov 23 07:36:59 crc kubenswrapper[4559]: I1123 07:36:59.430015 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-bf4c6585d-9tvg2_5b70e018-1ab3-4f23-b43f-1f433a096b81/manager/0.log" Nov 23 07:36:59 crc kubenswrapper[4559]: I1123 07:36:59.543126 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d86b44686-c776j_4cc947bb-881d-4592-9db6-ea53f6694ea5/kube-rbac-proxy/0.log" Nov 23 07:36:59 crc kubenswrapper[4559]: I1123 07:36:59.596374 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d86b44686-c776j_4cc947bb-881d-4592-9db6-ea53f6694ea5/manager/0.log" Nov 23 07:36:59 crc kubenswrapper[4559]: I1123 07:36:59.669820 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-769d9c7585-qq8sm_82279b3c-bca1-4891-b7d9-a367005ad84e/kube-rbac-proxy/0.log" Nov 23 07:36:59 crc kubenswrapper[4559]: I1123 07:36:59.748614 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5c75d7c94b-5tcjp_527a2a9e-741e-4f1b-8546-a852d73a836b/kube-rbac-proxy/0.log" Nov 23 07:36:59 crc kubenswrapper[4559]: I1123 07:36:59.830101 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-769d9c7585-qq8sm_82279b3c-bca1-4891-b7d9-a367005ad84e/manager/0.log" Nov 23 07:36:59 crc kubenswrapper[4559]: I1123 07:36:59.872775 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5c75d7c94b-5tcjp_527a2a9e-741e-4f1b-8546-a852d73a836b/manager/0.log" Nov 23 07:36:59 crc kubenswrapper[4559]: I1123 07:36:59.899983 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7879fb76fd-rxpxw_6935f82b-929a-421a-95d0-3315c9d0ef70/kube-rbac-proxy/0.log" Nov 23 07:37:00 crc kubenswrapper[4559]: I1123 07:37:00.035794 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7bb88cb858-qgvkm_50ff7c3e-8c4a-45e9-abe0-f5391a5c7c95/kube-rbac-proxy/0.log" Nov 23 07:37:00 crc kubenswrapper[4559]: I1123 07:37:00.055434 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7879fb76fd-rxpxw_6935f82b-929a-421a-95d0-3315c9d0ef70/manager/0.log" Nov 23 07:37:00 crc kubenswrapper[4559]: I1123 07:37:00.073124 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7bb88cb858-qgvkm_50ff7c3e-8c4a-45e9-abe0-f5391a5c7c95/manager/0.log" Nov 23 07:37:00 crc kubenswrapper[4559]: I1123 07:37:00.197862 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6f8c5b86cb-5rhv2_620fe37e-07bb-4286-b1e8-fd62db1cf022/kube-rbac-proxy/0.log" Nov 23 07:37:00 crc kubenswrapper[4559]: I1123 07:37:00.215025 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6f8c5b86cb-5rhv2_620fe37e-07bb-4286-b1e8-fd62db1cf022/manager/0.log" Nov 23 07:37:00 crc kubenswrapper[4559]: I1123 07:37:00.335791 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-66b7d6f598-hgbh8_7c53dd20-9115-4fac-a49e-b6b099150245/kube-rbac-proxy/0.log" Nov 23 07:37:00 crc kubenswrapper[4559]: I1123 07:37:00.397113 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-86d796d84d-6ht8s_38f7bf7e-6d60-4cda-a988-ef9ce22883f8/kube-rbac-proxy/0.log" Nov 23 07:37:00 crc kubenswrapper[4559]: I1123 07:37:00.399902 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-66b7d6f598-hgbh8_7c53dd20-9115-4fac-a49e-b6b099150245/manager/0.log" Nov 23 07:37:00 crc kubenswrapper[4559]: I1123 07:37:00.520975 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-86d796d84d-6ht8s_38f7bf7e-6d60-4cda-a988-ef9ce22883f8/manager/0.log" Nov 23 07:37:00 crc kubenswrapper[4559]: I1123 07:37:00.551144 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6fdc856c5d-bg99z_a659f54b-26d9-4681-963a-40348401f023/kube-rbac-proxy/0.log" Nov 23 07:37:00 crc kubenswrapper[4559]: I1123 07:37:00.599537 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6fdc856c5d-bg99z_a659f54b-26d9-4681-963a-40348401f023/manager/0.log" Nov 23 07:37:00 crc kubenswrapper[4559]: I1123 07:37:00.697003 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r_14445d59-81da-4b26-bd5d-2bbe2bd1a9c0/kube-rbac-proxy/0.log" Nov 23 07:37:00 crc kubenswrapper[4559]: I1123 07:37:00.712772 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-79d88dcd44pkg9r_14445d59-81da-4b26-bd5d-2bbe2bd1a9c0/manager/0.log" Nov 23 07:37:00 crc kubenswrapper[4559]: I1123 07:37:00.803022 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6cb9dc54f8-gmv99_ec988465-9f4e-46fe-beb3-e9032bf589d0/kube-rbac-proxy/0.log" Nov 23 07:37:00 crc kubenswrapper[4559]: I1123 07:37:00.916577 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-8486c7f98b-tfq8j_febcfb0d-e692-4e93-bbb5-40efe2929587/kube-rbac-proxy/0.log" Nov 23 07:37:01 crc kubenswrapper[4559]: I1123 07:37:01.113959 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-8486c7f98b-tfq8j_febcfb0d-e692-4e93-bbb5-40efe2929587/operator/0.log" Nov 23 07:37:01 crc kubenswrapper[4559]: I1123 07:37:01.192152 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-clkrn_7d86f9e5-a7b8-4807-bdeb-149b868a3c7a/registry-server/0.log" Nov 23 07:37:01 crc kubenswrapper[4559]: I1123 07:37:01.273244 4559 scope.go:117] "RemoveContainer" containerID="c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4" Nov 23 07:37:01 crc kubenswrapper[4559]: E1123 07:37:01.273541 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:37:01 crc kubenswrapper[4559]: I1123 07:37:01.288344 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5bdf4f7f7f-8hrnp_9c496b90-00da-4367-90f5-347a0c5a8ac6/kube-rbac-proxy/0.log" Nov 23 07:37:01 crc kubenswrapper[4559]: I1123 07:37:01.420602 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-6dc664666c-htkt7_c5127c1b-7342-4669-b04b-8484fd2ac326/kube-rbac-proxy/0.log" Nov 23 07:37:01 crc kubenswrapper[4559]: I1123 07:37:01.502985 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5bdf4f7f7f-8hrnp_9c496b90-00da-4367-90f5-347a0c5a8ac6/manager/0.log" Nov 23 07:37:01 crc kubenswrapper[4559]: I1123 07:37:01.540199 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-6dc664666c-htkt7_c5127c1b-7342-4669-b04b-8484fd2ac326/manager/0.log" Nov 23 07:37:01 crc kubenswrapper[4559]: I1123 07:37:01.656545 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-f6s9r_465518b5-58ac-4347-aae0-94b8f68a6b36/operator/0.log" Nov 23 07:37:01 crc kubenswrapper[4559]: I1123 07:37:01.695323 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6cb9dc54f8-gmv99_ec988465-9f4e-46fe-beb3-e9032bf589d0/manager/0.log" Nov 23 07:37:01 crc kubenswrapper[4559]: I1123 07:37:01.750448 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-799cb6ffd6-56qvk_b5998049-f03f-4de4-9601-dd970e3215fe/kube-rbac-proxy/0.log" Nov 23 07:37:01 crc kubenswrapper[4559]: I1123 07:37:01.836540 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-799cb6ffd6-56qvk_b5998049-f03f-4de4-9601-dd970e3215fe/manager/0.log" Nov 23 07:37:01 crc kubenswrapper[4559]: I1123 07:37:01.851765 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7798859c74-62tqn_6486d960-766f-4622-bad1-853e2b6a3fce/kube-rbac-proxy/0.log" Nov 23 07:37:01 crc kubenswrapper[4559]: I1123 07:37:01.939565 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7798859c74-62tqn_6486d960-766f-4622-bad1-853e2b6a3fce/manager/0.log" Nov 23 07:37:02 crc kubenswrapper[4559]: I1123 07:37:02.027313 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-8464cf66df-c2gjt_2463fe89-bc43-469f-9837-3ffd4b75605c/kube-rbac-proxy/0.log" Nov 23 07:37:02 crc kubenswrapper[4559]: I1123 07:37:02.047046 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-8464cf66df-c2gjt_2463fe89-bc43-469f-9837-3ffd4b75605c/manager/0.log" Nov 23 07:37:02 crc kubenswrapper[4559]: I1123 07:37:02.150874 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7cd4fb6f79-7brkl_a0f57f27-60d8-4fb8-ba76-c4ea5f3b99f6/kube-rbac-proxy/0.log" Nov 23 07:37:02 crc kubenswrapper[4559]: I1123 07:37:02.152319 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7cd4fb6f79-7brkl_a0f57f27-60d8-4fb8-ba76-c4ea5f3b99f6/manager/0.log" Nov 23 07:37:13 crc kubenswrapper[4559]: I1123 07:37:13.826865 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-2j2z7_ae746e9c-3187-4a3b-a439-b5ef25b37caf/control-plane-machine-set-operator/0.log" Nov 23 07:37:13 crc kubenswrapper[4559]: I1123 07:37:13.976603 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-ztnsp_18f99d77-1c2d-412b-93ae-1d1fc52f24ab/kube-rbac-proxy/0.log" Nov 23 07:37:13 crc kubenswrapper[4559]: I1123 07:37:13.999515 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-ztnsp_18f99d77-1c2d-412b-93ae-1d1fc52f24ab/machine-api-operator/0.log" Nov 23 07:37:14 crc kubenswrapper[4559]: I1123 07:37:14.273593 4559 scope.go:117] "RemoveContainer" containerID="c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4" Nov 23 07:37:14 crc kubenswrapper[4559]: E1123 07:37:14.273958 4559 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-s7zkt_openshift-machine-config-operator(4731beee-0cac-4189-8a70-743b0b709095)\"" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" Nov 23 07:37:23 crc kubenswrapper[4559]: I1123 07:37:23.988195 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-mqxbb_e5b5c7a9-2ccb-44ac-a6bd-826ffa8d7bda/cert-manager-controller/0.log" Nov 23 07:37:24 crc kubenswrapper[4559]: I1123 07:37:24.283474 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-qxxh5_74a9e3f9-a1b2-4f28-9ff4-a6d8a5d793b7/cert-manager-webhook/0.log" Nov 23 07:37:24 crc kubenswrapper[4559]: I1123 07:37:24.307190 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-m6h87_eb252f53-26dd-4071-8010-8a824926a679/cert-manager-cainjector/0.log" Nov 23 07:37:27 crc kubenswrapper[4559]: I1123 07:37:27.274107 4559 scope.go:117] "RemoveContainer" containerID="c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4" Nov 23 07:37:28 crc kubenswrapper[4559]: I1123 07:37:28.177767 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerStarted","Data":"60cceb458676c017283709cce5e12c3ed1c0a287b14502d52f9c3448758a8ab5"} Nov 23 07:37:33 crc kubenswrapper[4559]: I1123 07:37:33.349746 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-6qvx9_b0662ab8-3035-4b7b-b615-15cc0b963af4/nmstate-console-plugin/0.log" Nov 23 07:37:33 crc kubenswrapper[4559]: I1123 07:37:33.485455 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-fglqj_092f86a7-8e24-4a70-9af6-a4265c3688e5/nmstate-handler/0.log" Nov 23 07:37:33 crc kubenswrapper[4559]: I1123 07:37:33.525697 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-rpd7c_3755e847-69fa-47e4-93f9-fe15df377011/kube-rbac-proxy/0.log" Nov 23 07:37:33 crc kubenswrapper[4559]: I1123 07:37:33.554014 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-rpd7c_3755e847-69fa-47e4-93f9-fe15df377011/nmstate-metrics/0.log" Nov 23 07:37:33 crc kubenswrapper[4559]: I1123 07:37:33.633111 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-pfk9b_c7cb504c-5b1d-4d00-b5e9-6bcc6968ee84/nmstate-operator/0.log" Nov 23 07:37:33 crc kubenswrapper[4559]: I1123 07:37:33.723840 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-hv2js_6ff08b99-2dc5-4b9f-bef5-b0d34e76cb49/nmstate-webhook/0.log" Nov 23 07:37:43 crc kubenswrapper[4559]: I1123 07:37:43.194750 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-tgxdj_24dffaef-1c72-43f9-a5a8-ee0797a63077/kube-rbac-proxy/0.log" Nov 23 07:37:43 crc kubenswrapper[4559]: I1123 07:37:43.241784 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-tgxdj_24dffaef-1c72-43f9-a5a8-ee0797a63077/controller/0.log" Nov 23 07:37:43 crc kubenswrapper[4559]: I1123 07:37:43.351664 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/cp-frr-files/0.log" Nov 23 07:37:43 crc kubenswrapper[4559]: I1123 07:37:43.484033 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/cp-frr-files/0.log" Nov 23 07:37:43 crc kubenswrapper[4559]: I1123 07:37:43.485340 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/cp-metrics/0.log" Nov 23 07:37:43 crc kubenswrapper[4559]: I1123 07:37:43.491235 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/cp-reloader/0.log" Nov 23 07:37:43 crc kubenswrapper[4559]: I1123 07:37:43.524181 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/cp-reloader/0.log" Nov 23 07:37:43 crc kubenswrapper[4559]: I1123 07:37:43.609625 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/cp-frr-files/0.log" Nov 23 07:37:43 crc kubenswrapper[4559]: I1123 07:37:43.647722 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/cp-metrics/0.log" Nov 23 07:37:43 crc kubenswrapper[4559]: I1123 07:37:43.648713 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/cp-reloader/0.log" Nov 23 07:37:43 crc kubenswrapper[4559]: I1123 07:37:43.651066 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/cp-metrics/0.log" Nov 23 07:37:43 crc kubenswrapper[4559]: I1123 07:37:43.777711 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/cp-reloader/0.log" Nov 23 07:37:43 crc kubenswrapper[4559]: I1123 07:37:43.785303 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/cp-frr-files/0.log" Nov 23 07:37:43 crc kubenswrapper[4559]: I1123 07:37:43.788810 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/cp-metrics/0.log" Nov 23 07:37:43 crc kubenswrapper[4559]: I1123 07:37:43.824603 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/controller/0.log" Nov 23 07:37:43 crc kubenswrapper[4559]: I1123 07:37:43.921255 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/kube-rbac-proxy/0.log" Nov 23 07:37:43 crc kubenswrapper[4559]: I1123 07:37:43.931055 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/frr-metrics/0.log" Nov 23 07:37:43 crc kubenswrapper[4559]: I1123 07:37:43.957071 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/kube-rbac-proxy-frr/0.log" Nov 23 07:37:44 crc kubenswrapper[4559]: I1123 07:37:44.091109 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-zwlxs_5533445d-9129-4b9d-b87a-746df3caefb9/frr-k8s-webhook-server/0.log" Nov 23 07:37:44 crc kubenswrapper[4559]: I1123 07:37:44.091910 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/reloader/0.log" Nov 23 07:37:44 crc kubenswrapper[4559]: I1123 07:37:44.234156 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5c8ccb79b7-ngqhc_e57be21d-6409-4c23-b745-ec26bdf7e98d/manager/0.log" Nov 23 07:37:44 crc kubenswrapper[4559]: I1123 07:37:44.434291 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6bc96cc64c-zfqmv_85db5798-c470-4ec9-91c3-c9bdd9db327d/webhook-server/0.log" Nov 23 07:37:44 crc kubenswrapper[4559]: I1123 07:37:44.446348 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-stnln_fe556537-f24f-4a84-9f79-488938b3f4b5/kube-rbac-proxy/0.log" Nov 23 07:37:44 crc kubenswrapper[4559]: I1123 07:37:44.881376 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-stnln_fe556537-f24f-4a84-9f79-488938b3f4b5/speaker/0.log" Nov 23 07:37:45 crc kubenswrapper[4559]: I1123 07:37:45.011846 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-h2xjx_b3f1ab44-d9d7-4005-82b1-249213ea6ee6/frr/0.log" Nov 23 07:37:52 crc kubenswrapper[4559]: I1123 07:37:52.970054 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg_44c771c2-63aa-45f3-8551-f6b127f6121c/util/0.log" Nov 23 07:37:53 crc kubenswrapper[4559]: I1123 07:37:53.124948 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg_44c771c2-63aa-45f3-8551-f6b127f6121c/pull/0.log" Nov 23 07:37:53 crc kubenswrapper[4559]: I1123 07:37:53.139091 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg_44c771c2-63aa-45f3-8551-f6b127f6121c/util/0.log" Nov 23 07:37:53 crc kubenswrapper[4559]: I1123 07:37:53.148993 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg_44c771c2-63aa-45f3-8551-f6b127f6121c/pull/0.log" Nov 23 07:37:53 crc kubenswrapper[4559]: I1123 07:37:53.271512 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg_44c771c2-63aa-45f3-8551-f6b127f6121c/util/0.log" Nov 23 07:37:53 crc kubenswrapper[4559]: I1123 07:37:53.294009 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg_44c771c2-63aa-45f3-8551-f6b127f6121c/pull/0.log" Nov 23 07:37:53 crc kubenswrapper[4559]: I1123 07:37:53.318921 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772egrtlg_44c771c2-63aa-45f3-8551-f6b127f6121c/extract/0.log" Nov 23 07:37:53 crc kubenswrapper[4559]: I1123 07:37:53.418184 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nc2rk_4d8f5bd0-c208-43fb-acd6-496d94f6dec3/extract-utilities/0.log" Nov 23 07:37:53 crc kubenswrapper[4559]: I1123 07:37:53.543322 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nc2rk_4d8f5bd0-c208-43fb-acd6-496d94f6dec3/extract-utilities/0.log" Nov 23 07:37:53 crc kubenswrapper[4559]: I1123 07:37:53.547580 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nc2rk_4d8f5bd0-c208-43fb-acd6-496d94f6dec3/extract-content/0.log" Nov 23 07:37:53 crc kubenswrapper[4559]: I1123 07:37:53.552147 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nc2rk_4d8f5bd0-c208-43fb-acd6-496d94f6dec3/extract-content/0.log" Nov 23 07:37:53 crc kubenswrapper[4559]: I1123 07:37:53.664704 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nc2rk_4d8f5bd0-c208-43fb-acd6-496d94f6dec3/extract-utilities/0.log" Nov 23 07:37:53 crc kubenswrapper[4559]: I1123 07:37:53.714678 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nc2rk_4d8f5bd0-c208-43fb-acd6-496d94f6dec3/extract-content/0.log" Nov 23 07:37:53 crc kubenswrapper[4559]: I1123 07:37:53.826268 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nw54h_c2daeb5e-241f-4764-962b-64c9c3d34597/extract-utilities/0.log" Nov 23 07:37:54 crc kubenswrapper[4559]: I1123 07:37:54.005613 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nw54h_c2daeb5e-241f-4764-962b-64c9c3d34597/extract-content/0.log" Nov 23 07:37:54 crc kubenswrapper[4559]: I1123 07:37:54.016953 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nw54h_c2daeb5e-241f-4764-962b-64c9c3d34597/extract-utilities/0.log" Nov 23 07:37:54 crc kubenswrapper[4559]: I1123 07:37:54.041787 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nw54h_c2daeb5e-241f-4764-962b-64c9c3d34597/extract-content/0.log" Nov 23 07:37:54 crc kubenswrapper[4559]: I1123 07:37:54.077633 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nc2rk_4d8f5bd0-c208-43fb-acd6-496d94f6dec3/registry-server/0.log" Nov 23 07:37:54 crc kubenswrapper[4559]: I1123 07:37:54.182117 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nw54h_c2daeb5e-241f-4764-962b-64c9c3d34597/extract-content/0.log" Nov 23 07:37:54 crc kubenswrapper[4559]: I1123 07:37:54.223776 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nw54h_c2daeb5e-241f-4764-962b-64c9c3d34597/extract-utilities/0.log" Nov 23 07:37:54 crc kubenswrapper[4559]: I1123 07:37:54.333695 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nw54h_c2daeb5e-241f-4764-962b-64c9c3d34597/registry-server/0.log" Nov 23 07:37:54 crc kubenswrapper[4559]: I1123 07:37:54.359343 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458_23f858ff-40d3-4f2a-ac00-f7e3b8c544f1/util/0.log" Nov 23 07:37:54 crc kubenswrapper[4559]: I1123 07:37:54.500357 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458_23f858ff-40d3-4f2a-ac00-f7e3b8c544f1/util/0.log" Nov 23 07:37:54 crc kubenswrapper[4559]: I1123 07:37:54.517888 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458_23f858ff-40d3-4f2a-ac00-f7e3b8c544f1/pull/0.log" Nov 23 07:37:54 crc kubenswrapper[4559]: I1123 07:37:54.518186 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458_23f858ff-40d3-4f2a-ac00-f7e3b8c544f1/pull/0.log" Nov 23 07:37:54 crc kubenswrapper[4559]: I1123 07:37:54.629715 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458_23f858ff-40d3-4f2a-ac00-f7e3b8c544f1/util/0.log" Nov 23 07:37:54 crc kubenswrapper[4559]: I1123 07:37:54.647125 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458_23f858ff-40d3-4f2a-ac00-f7e3b8c544f1/extract/0.log" Nov 23 07:37:54 crc kubenswrapper[4559]: I1123 07:37:54.669847 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c64g458_23f858ff-40d3-4f2a-ac00-f7e3b8c544f1/pull/0.log" Nov 23 07:37:54 crc kubenswrapper[4559]: I1123 07:37:54.767530 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-v7hpt_351f8b04-cba3-4484-94b6-c4abd1c9be42/marketplace-operator/0.log" Nov 23 07:37:54 crc kubenswrapper[4559]: I1123 07:37:54.800694 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4zrs4_b5a957ce-5a85-497c-9794-05ac9b6c92ae/extract-utilities/0.log" Nov 23 07:37:54 crc kubenswrapper[4559]: I1123 07:37:54.965783 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4zrs4_b5a957ce-5a85-497c-9794-05ac9b6c92ae/extract-content/0.log" Nov 23 07:37:54 crc kubenswrapper[4559]: I1123 07:37:54.970282 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4zrs4_b5a957ce-5a85-497c-9794-05ac9b6c92ae/extract-utilities/0.log" Nov 23 07:37:54 crc kubenswrapper[4559]: I1123 07:37:54.985045 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4zrs4_b5a957ce-5a85-497c-9794-05ac9b6c92ae/extract-content/0.log" Nov 23 07:37:55 crc kubenswrapper[4559]: I1123 07:37:55.088683 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4zrs4_b5a957ce-5a85-497c-9794-05ac9b6c92ae/extract-content/0.log" Nov 23 07:37:55 crc kubenswrapper[4559]: I1123 07:37:55.089375 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4zrs4_b5a957ce-5a85-497c-9794-05ac9b6c92ae/extract-utilities/0.log" Nov 23 07:37:55 crc kubenswrapper[4559]: I1123 07:37:55.196281 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4zrs4_b5a957ce-5a85-497c-9794-05ac9b6c92ae/registry-server/0.log" Nov 23 07:37:55 crc kubenswrapper[4559]: I1123 07:37:55.243851 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w2mts_0579529b-d4e5-458c-a2d6-9d0d7026c7b8/extract-utilities/0.log" Nov 23 07:37:55 crc kubenswrapper[4559]: I1123 07:37:55.336483 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w2mts_0579529b-d4e5-458c-a2d6-9d0d7026c7b8/extract-content/0.log" Nov 23 07:37:55 crc kubenswrapper[4559]: I1123 07:37:55.362203 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w2mts_0579529b-d4e5-458c-a2d6-9d0d7026c7b8/extract-utilities/0.log" Nov 23 07:37:55 crc kubenswrapper[4559]: I1123 07:37:55.373482 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w2mts_0579529b-d4e5-458c-a2d6-9d0d7026c7b8/extract-content/0.log" Nov 23 07:37:55 crc kubenswrapper[4559]: I1123 07:37:55.484679 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w2mts_0579529b-d4e5-458c-a2d6-9d0d7026c7b8/extract-content/0.log" Nov 23 07:37:55 crc kubenswrapper[4559]: I1123 07:37:55.516375 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w2mts_0579529b-d4e5-458c-a2d6-9d0d7026c7b8/extract-utilities/0.log" Nov 23 07:37:55 crc kubenswrapper[4559]: I1123 07:37:55.807286 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w2mts_0579529b-d4e5-458c-a2d6-9d0d7026c7b8/registry-server/0.log" Nov 23 07:39:07 crc kubenswrapper[4559]: I1123 07:39:07.859493 4559 generic.go:334] "Generic (PLEG): container finished" podID="08ac5460-3a47-487d-9b86-93c6debd86d9" containerID="7b6a158235ebef33f14538cf8e75beb0fadadd55a9ef4c3fb415fa7c9a93e5a2" exitCode=0 Nov 23 07:39:07 crc kubenswrapper[4559]: I1123 07:39:07.859568 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m5msz/must-gather-fgqp7" event={"ID":"08ac5460-3a47-487d-9b86-93c6debd86d9","Type":"ContainerDied","Data":"7b6a158235ebef33f14538cf8e75beb0fadadd55a9ef4c3fb415fa7c9a93e5a2"} Nov 23 07:39:07 crc kubenswrapper[4559]: I1123 07:39:07.860445 4559 scope.go:117] "RemoveContainer" containerID="7b6a158235ebef33f14538cf8e75beb0fadadd55a9ef4c3fb415fa7c9a93e5a2" Nov 23 07:39:08 crc kubenswrapper[4559]: I1123 07:39:08.075357 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-m5msz_must-gather-fgqp7_08ac5460-3a47-487d-9b86-93c6debd86d9/gather/0.log" Nov 23 07:39:16 crc kubenswrapper[4559]: I1123 07:39:16.582166 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-m5msz/must-gather-fgqp7"] Nov 23 07:39:16 crc kubenswrapper[4559]: I1123 07:39:16.582807 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-m5msz/must-gather-fgqp7" podUID="08ac5460-3a47-487d-9b86-93c6debd86d9" containerName="copy" containerID="cri-o://abeb131c06d46e0814ddc07e00b54f898ca86cf245e8912002af75411617152c" gracePeriod=2 Nov 23 07:39:16 crc kubenswrapper[4559]: I1123 07:39:16.593380 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-m5msz/must-gather-fgqp7"] Nov 23 07:39:16 crc kubenswrapper[4559]: I1123 07:39:16.920987 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-m5msz_must-gather-fgqp7_08ac5460-3a47-487d-9b86-93c6debd86d9/copy/0.log" Nov 23 07:39:16 crc kubenswrapper[4559]: I1123 07:39:16.921819 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m5msz/must-gather-fgqp7" Nov 23 07:39:16 crc kubenswrapper[4559]: I1123 07:39:16.931309 4559 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-m5msz_must-gather-fgqp7_08ac5460-3a47-487d-9b86-93c6debd86d9/copy/0.log" Nov 23 07:39:16 crc kubenswrapper[4559]: I1123 07:39:16.931674 4559 generic.go:334] "Generic (PLEG): container finished" podID="08ac5460-3a47-487d-9b86-93c6debd86d9" containerID="abeb131c06d46e0814ddc07e00b54f898ca86cf245e8912002af75411617152c" exitCode=143 Nov 23 07:39:16 crc kubenswrapper[4559]: I1123 07:39:16.931729 4559 scope.go:117] "RemoveContainer" containerID="abeb131c06d46e0814ddc07e00b54f898ca86cf245e8912002af75411617152c" Nov 23 07:39:16 crc kubenswrapper[4559]: I1123 07:39:16.931740 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m5msz/must-gather-fgqp7" Nov 23 07:39:16 crc kubenswrapper[4559]: I1123 07:39:16.952910 4559 scope.go:117] "RemoveContainer" containerID="7b6a158235ebef33f14538cf8e75beb0fadadd55a9ef4c3fb415fa7c9a93e5a2" Nov 23 07:39:16 crc kubenswrapper[4559]: I1123 07:39:16.953910 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/08ac5460-3a47-487d-9b86-93c6debd86d9-must-gather-output\") pod \"08ac5460-3a47-487d-9b86-93c6debd86d9\" (UID: \"08ac5460-3a47-487d-9b86-93c6debd86d9\") " Nov 23 07:39:16 crc kubenswrapper[4559]: I1123 07:39:16.953981 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbxts\" (UniqueName: \"kubernetes.io/projected/08ac5460-3a47-487d-9b86-93c6debd86d9-kube-api-access-gbxts\") pod \"08ac5460-3a47-487d-9b86-93c6debd86d9\" (UID: \"08ac5460-3a47-487d-9b86-93c6debd86d9\") " Nov 23 07:39:16 crc kubenswrapper[4559]: I1123 07:39:16.961236 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08ac5460-3a47-487d-9b86-93c6debd86d9-kube-api-access-gbxts" (OuterVolumeSpecName: "kube-api-access-gbxts") pod "08ac5460-3a47-487d-9b86-93c6debd86d9" (UID: "08ac5460-3a47-487d-9b86-93c6debd86d9"). InnerVolumeSpecName "kube-api-access-gbxts". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:39:17 crc kubenswrapper[4559]: I1123 07:39:17.034349 4559 scope.go:117] "RemoveContainer" containerID="abeb131c06d46e0814ddc07e00b54f898ca86cf245e8912002af75411617152c" Nov 23 07:39:17 crc kubenswrapper[4559]: E1123 07:39:17.035021 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abeb131c06d46e0814ddc07e00b54f898ca86cf245e8912002af75411617152c\": container with ID starting with abeb131c06d46e0814ddc07e00b54f898ca86cf245e8912002af75411617152c not found: ID does not exist" containerID="abeb131c06d46e0814ddc07e00b54f898ca86cf245e8912002af75411617152c" Nov 23 07:39:17 crc kubenswrapper[4559]: I1123 07:39:17.035058 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abeb131c06d46e0814ddc07e00b54f898ca86cf245e8912002af75411617152c"} err="failed to get container status \"abeb131c06d46e0814ddc07e00b54f898ca86cf245e8912002af75411617152c\": rpc error: code = NotFound desc = could not find container \"abeb131c06d46e0814ddc07e00b54f898ca86cf245e8912002af75411617152c\": container with ID starting with abeb131c06d46e0814ddc07e00b54f898ca86cf245e8912002af75411617152c not found: ID does not exist" Nov 23 07:39:17 crc kubenswrapper[4559]: I1123 07:39:17.035086 4559 scope.go:117] "RemoveContainer" containerID="7b6a158235ebef33f14538cf8e75beb0fadadd55a9ef4c3fb415fa7c9a93e5a2" Nov 23 07:39:17 crc kubenswrapper[4559]: E1123 07:39:17.039952 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b6a158235ebef33f14538cf8e75beb0fadadd55a9ef4c3fb415fa7c9a93e5a2\": container with ID starting with 7b6a158235ebef33f14538cf8e75beb0fadadd55a9ef4c3fb415fa7c9a93e5a2 not found: ID does not exist" containerID="7b6a158235ebef33f14538cf8e75beb0fadadd55a9ef4c3fb415fa7c9a93e5a2" Nov 23 07:39:17 crc kubenswrapper[4559]: I1123 07:39:17.039981 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b6a158235ebef33f14538cf8e75beb0fadadd55a9ef4c3fb415fa7c9a93e5a2"} err="failed to get container status \"7b6a158235ebef33f14538cf8e75beb0fadadd55a9ef4c3fb415fa7c9a93e5a2\": rpc error: code = NotFound desc = could not find container \"7b6a158235ebef33f14538cf8e75beb0fadadd55a9ef4c3fb415fa7c9a93e5a2\": container with ID starting with 7b6a158235ebef33f14538cf8e75beb0fadadd55a9ef4c3fb415fa7c9a93e5a2 not found: ID does not exist" Nov 23 07:39:17 crc kubenswrapper[4559]: I1123 07:39:17.056146 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbxts\" (UniqueName: \"kubernetes.io/projected/08ac5460-3a47-487d-9b86-93c6debd86d9-kube-api-access-gbxts\") on node \"crc\" DevicePath \"\"" Nov 23 07:39:17 crc kubenswrapper[4559]: I1123 07:39:17.118132 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08ac5460-3a47-487d-9b86-93c6debd86d9-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "08ac5460-3a47-487d-9b86-93c6debd86d9" (UID: "08ac5460-3a47-487d-9b86-93c6debd86d9"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:39:17 crc kubenswrapper[4559]: I1123 07:39:17.157819 4559 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/08ac5460-3a47-487d-9b86-93c6debd86d9-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 23 07:39:18 crc kubenswrapper[4559]: I1123 07:39:18.281000 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08ac5460-3a47-487d-9b86-93c6debd86d9" path="/var/lib/kubelet/pods/08ac5460-3a47-487d-9b86-93c6debd86d9/volumes" Nov 23 07:39:56 crc kubenswrapper[4559]: I1123 07:39:56.166564 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:39:56 crc kubenswrapper[4559]: I1123 07:39:56.166974 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:40:14 crc kubenswrapper[4559]: I1123 07:40:14.597218 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-6b4fdd88f9-9hnng" podUID="db083799-cb94-42d3-8a54-8446f7a76502" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Nov 23 07:40:26 crc kubenswrapper[4559]: I1123 07:40:26.167568 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:40:26 crc kubenswrapper[4559]: I1123 07:40:26.167959 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:40:56 crc kubenswrapper[4559]: I1123 07:40:56.167362 4559 patch_prober.go:28] interesting pod/machine-config-daemon-s7zkt container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 07:40:56 crc kubenswrapper[4559]: I1123 07:40:56.167793 4559 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 07:40:56 crc kubenswrapper[4559]: I1123 07:40:56.167835 4559 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" Nov 23 07:40:56 crc kubenswrapper[4559]: I1123 07:40:56.168501 4559 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"60cceb458676c017283709cce5e12c3ed1c0a287b14502d52f9c3448758a8ab5"} pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 07:40:56 crc kubenswrapper[4559]: I1123 07:40:56.168550 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" podUID="4731beee-0cac-4189-8a70-743b0b709095" containerName="machine-config-daemon" containerID="cri-o://60cceb458676c017283709cce5e12c3ed1c0a287b14502d52f9c3448758a8ab5" gracePeriod=600 Nov 23 07:40:56 crc kubenswrapper[4559]: I1123 07:40:56.550445 4559 generic.go:334] "Generic (PLEG): container finished" podID="4731beee-0cac-4189-8a70-743b0b709095" containerID="60cceb458676c017283709cce5e12c3ed1c0a287b14502d52f9c3448758a8ab5" exitCode=0 Nov 23 07:40:56 crc kubenswrapper[4559]: I1123 07:40:56.550510 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerDied","Data":"60cceb458676c017283709cce5e12c3ed1c0a287b14502d52f9c3448758a8ab5"} Nov 23 07:40:56 crc kubenswrapper[4559]: I1123 07:40:56.550715 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-s7zkt" event={"ID":"4731beee-0cac-4189-8a70-743b0b709095","Type":"ContainerStarted","Data":"b2a227f2649183bea84130f256f6cc7cbaf6137bba7b1a91cd8040dc3597abea"} Nov 23 07:40:56 crc kubenswrapper[4559]: I1123 07:40:56.550734 4559 scope.go:117] "RemoveContainer" containerID="c6ededc9d40d41fe408609ef1376311487c7f8a1c83072b77eac2bf9f357c7b4" Nov 23 07:41:00 crc kubenswrapper[4559]: I1123 07:41:00.380658 4559 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gs58d"] Nov 23 07:41:00 crc kubenswrapper[4559]: E1123 07:41:00.381355 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08ac5460-3a47-487d-9b86-93c6debd86d9" containerName="gather" Nov 23 07:41:00 crc kubenswrapper[4559]: I1123 07:41:00.381369 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="08ac5460-3a47-487d-9b86-93c6debd86d9" containerName="gather" Nov 23 07:41:00 crc kubenswrapper[4559]: E1123 07:41:00.381389 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9b6f723-e379-4cfb-8327-ae8289d9cf73" containerName="container-00" Nov 23 07:41:00 crc kubenswrapper[4559]: I1123 07:41:00.381394 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9b6f723-e379-4cfb-8327-ae8289d9cf73" containerName="container-00" Nov 23 07:41:00 crc kubenswrapper[4559]: E1123 07:41:00.381413 4559 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08ac5460-3a47-487d-9b86-93c6debd86d9" containerName="copy" Nov 23 07:41:00 crc kubenswrapper[4559]: I1123 07:41:00.381418 4559 state_mem.go:107] "Deleted CPUSet assignment" podUID="08ac5460-3a47-487d-9b86-93c6debd86d9" containerName="copy" Nov 23 07:41:00 crc kubenswrapper[4559]: I1123 07:41:00.381565 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9b6f723-e379-4cfb-8327-ae8289d9cf73" containerName="container-00" Nov 23 07:41:00 crc kubenswrapper[4559]: I1123 07:41:00.381579 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="08ac5460-3a47-487d-9b86-93c6debd86d9" containerName="gather" Nov 23 07:41:00 crc kubenswrapper[4559]: I1123 07:41:00.381595 4559 memory_manager.go:354] "RemoveStaleState removing state" podUID="08ac5460-3a47-487d-9b86-93c6debd86d9" containerName="copy" Nov 23 07:41:00 crc kubenswrapper[4559]: I1123 07:41:00.382712 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gs58d" Nov 23 07:41:00 crc kubenswrapper[4559]: I1123 07:41:00.396564 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gs58d"] Nov 23 07:41:00 crc kubenswrapper[4559]: I1123 07:41:00.454998 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ws49w\" (UniqueName: \"kubernetes.io/projected/8db86fb4-90eb-472b-8540-fa7d61dc3677-kube-api-access-ws49w\") pod \"community-operators-gs58d\" (UID: \"8db86fb4-90eb-472b-8540-fa7d61dc3677\") " pod="openshift-marketplace/community-operators-gs58d" Nov 23 07:41:00 crc kubenswrapper[4559]: I1123 07:41:00.455176 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8db86fb4-90eb-472b-8540-fa7d61dc3677-utilities\") pod \"community-operators-gs58d\" (UID: \"8db86fb4-90eb-472b-8540-fa7d61dc3677\") " pod="openshift-marketplace/community-operators-gs58d" Nov 23 07:41:00 crc kubenswrapper[4559]: I1123 07:41:00.455237 4559 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8db86fb4-90eb-472b-8540-fa7d61dc3677-catalog-content\") pod \"community-operators-gs58d\" (UID: \"8db86fb4-90eb-472b-8540-fa7d61dc3677\") " pod="openshift-marketplace/community-operators-gs58d" Nov 23 07:41:00 crc kubenswrapper[4559]: I1123 07:41:00.556908 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8db86fb4-90eb-472b-8540-fa7d61dc3677-utilities\") pod \"community-operators-gs58d\" (UID: \"8db86fb4-90eb-472b-8540-fa7d61dc3677\") " pod="openshift-marketplace/community-operators-gs58d" Nov 23 07:41:00 crc kubenswrapper[4559]: I1123 07:41:00.556976 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8db86fb4-90eb-472b-8540-fa7d61dc3677-catalog-content\") pod \"community-operators-gs58d\" (UID: \"8db86fb4-90eb-472b-8540-fa7d61dc3677\") " pod="openshift-marketplace/community-operators-gs58d" Nov 23 07:41:00 crc kubenswrapper[4559]: I1123 07:41:00.557135 4559 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ws49w\" (UniqueName: \"kubernetes.io/projected/8db86fb4-90eb-472b-8540-fa7d61dc3677-kube-api-access-ws49w\") pod \"community-operators-gs58d\" (UID: \"8db86fb4-90eb-472b-8540-fa7d61dc3677\") " pod="openshift-marketplace/community-operators-gs58d" Nov 23 07:41:00 crc kubenswrapper[4559]: I1123 07:41:00.557361 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8db86fb4-90eb-472b-8540-fa7d61dc3677-utilities\") pod \"community-operators-gs58d\" (UID: \"8db86fb4-90eb-472b-8540-fa7d61dc3677\") " pod="openshift-marketplace/community-operators-gs58d" Nov 23 07:41:00 crc kubenswrapper[4559]: I1123 07:41:00.557404 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8db86fb4-90eb-472b-8540-fa7d61dc3677-catalog-content\") pod \"community-operators-gs58d\" (UID: \"8db86fb4-90eb-472b-8540-fa7d61dc3677\") " pod="openshift-marketplace/community-operators-gs58d" Nov 23 07:41:00 crc kubenswrapper[4559]: I1123 07:41:00.577102 4559 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ws49w\" (UniqueName: \"kubernetes.io/projected/8db86fb4-90eb-472b-8540-fa7d61dc3677-kube-api-access-ws49w\") pod \"community-operators-gs58d\" (UID: \"8db86fb4-90eb-472b-8540-fa7d61dc3677\") " pod="openshift-marketplace/community-operators-gs58d" Nov 23 07:41:00 crc kubenswrapper[4559]: I1123 07:41:00.700470 4559 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gs58d" Nov 23 07:41:01 crc kubenswrapper[4559]: I1123 07:41:01.089707 4559 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gs58d"] Nov 23 07:41:01 crc kubenswrapper[4559]: I1123 07:41:01.586758 4559 generic.go:334] "Generic (PLEG): container finished" podID="8db86fb4-90eb-472b-8540-fa7d61dc3677" containerID="e922e7c9ad2fdbe32e0ad79f1babd55970e7cc403abbd4632fad7c43796f0474" exitCode=0 Nov 23 07:41:01 crc kubenswrapper[4559]: I1123 07:41:01.586793 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gs58d" event={"ID":"8db86fb4-90eb-472b-8540-fa7d61dc3677","Type":"ContainerDied","Data":"e922e7c9ad2fdbe32e0ad79f1babd55970e7cc403abbd4632fad7c43796f0474"} Nov 23 07:41:01 crc kubenswrapper[4559]: I1123 07:41:01.586813 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gs58d" event={"ID":"8db86fb4-90eb-472b-8540-fa7d61dc3677","Type":"ContainerStarted","Data":"7a7b4be4747dc8d86cda6e2643439d0d4d7aa7bbfa7dd4c288f2e512d7c77171"} Nov 23 07:41:02 crc kubenswrapper[4559]: I1123 07:41:02.594324 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gs58d" event={"ID":"8db86fb4-90eb-472b-8540-fa7d61dc3677","Type":"ContainerStarted","Data":"e5b2c31e701ac4238b9694f1e091432b2530370472ef8bce022425afd72c4acd"} Nov 23 07:41:03 crc kubenswrapper[4559]: I1123 07:41:03.602805 4559 generic.go:334] "Generic (PLEG): container finished" podID="8db86fb4-90eb-472b-8540-fa7d61dc3677" containerID="e5b2c31e701ac4238b9694f1e091432b2530370472ef8bce022425afd72c4acd" exitCode=0 Nov 23 07:41:03 crc kubenswrapper[4559]: I1123 07:41:03.602845 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gs58d" event={"ID":"8db86fb4-90eb-472b-8540-fa7d61dc3677","Type":"ContainerDied","Data":"e5b2c31e701ac4238b9694f1e091432b2530370472ef8bce022425afd72c4acd"} Nov 23 07:41:03 crc kubenswrapper[4559]: I1123 07:41:03.605649 4559 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 07:41:04 crc kubenswrapper[4559]: I1123 07:41:04.610868 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gs58d" event={"ID":"8db86fb4-90eb-472b-8540-fa7d61dc3677","Type":"ContainerStarted","Data":"b335dfe4be46a9e194b0908bf05e0889b4c6e2a2304adcc980a5c3199a246d8c"} Nov 23 07:41:10 crc kubenswrapper[4559]: I1123 07:41:10.700818 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gs58d" Nov 23 07:41:10 crc kubenswrapper[4559]: I1123 07:41:10.701907 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gs58d" Nov 23 07:41:10 crc kubenswrapper[4559]: I1123 07:41:10.733516 4559 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gs58d" Nov 23 07:41:10 crc kubenswrapper[4559]: I1123 07:41:10.747318 4559 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gs58d" podStartSLOduration=8.27396359 podStartE2EDuration="10.747303607s" podCreationTimestamp="2025-11-23 07:41:00 +0000 UTC" firstStartedPulling="2025-11-23 07:41:01.587991962 +0000 UTC m=+3363.609977576" lastFinishedPulling="2025-11-23 07:41:04.061331979 +0000 UTC m=+3366.083317593" observedRunningTime="2025-11-23 07:41:04.630898657 +0000 UTC m=+3366.652884271" watchObservedRunningTime="2025-11-23 07:41:10.747303607 +0000 UTC m=+3372.769289221" Nov 23 07:41:11 crc kubenswrapper[4559]: I1123 07:41:11.683566 4559 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gs58d" Nov 23 07:41:11 crc kubenswrapper[4559]: I1123 07:41:11.715621 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gs58d"] Nov 23 07:41:13 crc kubenswrapper[4559]: I1123 07:41:13.666803 4559 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gs58d" podUID="8db86fb4-90eb-472b-8540-fa7d61dc3677" containerName="registry-server" containerID="cri-o://b335dfe4be46a9e194b0908bf05e0889b4c6e2a2304adcc980a5c3199a246d8c" gracePeriod=2 Nov 23 07:41:14 crc kubenswrapper[4559]: I1123 07:41:14.031183 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gs58d" Nov 23 07:41:14 crc kubenswrapper[4559]: I1123 07:41:14.177008 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ws49w\" (UniqueName: \"kubernetes.io/projected/8db86fb4-90eb-472b-8540-fa7d61dc3677-kube-api-access-ws49w\") pod \"8db86fb4-90eb-472b-8540-fa7d61dc3677\" (UID: \"8db86fb4-90eb-472b-8540-fa7d61dc3677\") " Nov 23 07:41:14 crc kubenswrapper[4559]: I1123 07:41:14.177241 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8db86fb4-90eb-472b-8540-fa7d61dc3677-catalog-content\") pod \"8db86fb4-90eb-472b-8540-fa7d61dc3677\" (UID: \"8db86fb4-90eb-472b-8540-fa7d61dc3677\") " Nov 23 07:41:14 crc kubenswrapper[4559]: I1123 07:41:14.177264 4559 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8db86fb4-90eb-472b-8540-fa7d61dc3677-utilities\") pod \"8db86fb4-90eb-472b-8540-fa7d61dc3677\" (UID: \"8db86fb4-90eb-472b-8540-fa7d61dc3677\") " Nov 23 07:41:14 crc kubenswrapper[4559]: I1123 07:41:14.177887 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8db86fb4-90eb-472b-8540-fa7d61dc3677-utilities" (OuterVolumeSpecName: "utilities") pod "8db86fb4-90eb-472b-8540-fa7d61dc3677" (UID: "8db86fb4-90eb-472b-8540-fa7d61dc3677"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:41:14 crc kubenswrapper[4559]: I1123 07:41:14.178298 4559 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8db86fb4-90eb-472b-8540-fa7d61dc3677-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 07:41:14 crc kubenswrapper[4559]: I1123 07:41:14.181676 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8db86fb4-90eb-472b-8540-fa7d61dc3677-kube-api-access-ws49w" (OuterVolumeSpecName: "kube-api-access-ws49w") pod "8db86fb4-90eb-472b-8540-fa7d61dc3677" (UID: "8db86fb4-90eb-472b-8540-fa7d61dc3677"). InnerVolumeSpecName "kube-api-access-ws49w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 07:41:14 crc kubenswrapper[4559]: I1123 07:41:14.216031 4559 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8db86fb4-90eb-472b-8540-fa7d61dc3677-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8db86fb4-90eb-472b-8540-fa7d61dc3677" (UID: "8db86fb4-90eb-472b-8540-fa7d61dc3677"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 07:41:14 crc kubenswrapper[4559]: I1123 07:41:14.279998 4559 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8db86fb4-90eb-472b-8540-fa7d61dc3677-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 07:41:14 crc kubenswrapper[4559]: I1123 07:41:14.280144 4559 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ws49w\" (UniqueName: \"kubernetes.io/projected/8db86fb4-90eb-472b-8540-fa7d61dc3677-kube-api-access-ws49w\") on node \"crc\" DevicePath \"\"" Nov 23 07:41:14 crc kubenswrapper[4559]: I1123 07:41:14.674554 4559 generic.go:334] "Generic (PLEG): container finished" podID="8db86fb4-90eb-472b-8540-fa7d61dc3677" containerID="b335dfe4be46a9e194b0908bf05e0889b4c6e2a2304adcc980a5c3199a246d8c" exitCode=0 Nov 23 07:41:14 crc kubenswrapper[4559]: I1123 07:41:14.674601 4559 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gs58d" Nov 23 07:41:14 crc kubenswrapper[4559]: I1123 07:41:14.674617 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gs58d" event={"ID":"8db86fb4-90eb-472b-8540-fa7d61dc3677","Type":"ContainerDied","Data":"b335dfe4be46a9e194b0908bf05e0889b4c6e2a2304adcc980a5c3199a246d8c"} Nov 23 07:41:14 crc kubenswrapper[4559]: I1123 07:41:14.674936 4559 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gs58d" event={"ID":"8db86fb4-90eb-472b-8540-fa7d61dc3677","Type":"ContainerDied","Data":"7a7b4be4747dc8d86cda6e2643439d0d4d7aa7bbfa7dd4c288f2e512d7c77171"} Nov 23 07:41:14 crc kubenswrapper[4559]: I1123 07:41:14.674983 4559 scope.go:117] "RemoveContainer" containerID="b335dfe4be46a9e194b0908bf05e0889b4c6e2a2304adcc980a5c3199a246d8c" Nov 23 07:41:14 crc kubenswrapper[4559]: I1123 07:41:14.692143 4559 scope.go:117] "RemoveContainer" containerID="e5b2c31e701ac4238b9694f1e091432b2530370472ef8bce022425afd72c4acd" Nov 23 07:41:14 crc kubenswrapper[4559]: I1123 07:41:14.692237 4559 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gs58d"] Nov 23 07:41:14 crc kubenswrapper[4559]: I1123 07:41:14.698817 4559 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gs58d"] Nov 23 07:41:14 crc kubenswrapper[4559]: I1123 07:41:14.711494 4559 scope.go:117] "RemoveContainer" containerID="e922e7c9ad2fdbe32e0ad79f1babd55970e7cc403abbd4632fad7c43796f0474" Nov 23 07:41:14 crc kubenswrapper[4559]: I1123 07:41:14.741264 4559 scope.go:117] "RemoveContainer" containerID="b335dfe4be46a9e194b0908bf05e0889b4c6e2a2304adcc980a5c3199a246d8c" Nov 23 07:41:14 crc kubenswrapper[4559]: E1123 07:41:14.741586 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b335dfe4be46a9e194b0908bf05e0889b4c6e2a2304adcc980a5c3199a246d8c\": container with ID starting with b335dfe4be46a9e194b0908bf05e0889b4c6e2a2304adcc980a5c3199a246d8c not found: ID does not exist" containerID="b335dfe4be46a9e194b0908bf05e0889b4c6e2a2304adcc980a5c3199a246d8c" Nov 23 07:41:14 crc kubenswrapper[4559]: I1123 07:41:14.741636 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b335dfe4be46a9e194b0908bf05e0889b4c6e2a2304adcc980a5c3199a246d8c"} err="failed to get container status \"b335dfe4be46a9e194b0908bf05e0889b4c6e2a2304adcc980a5c3199a246d8c\": rpc error: code = NotFound desc = could not find container \"b335dfe4be46a9e194b0908bf05e0889b4c6e2a2304adcc980a5c3199a246d8c\": container with ID starting with b335dfe4be46a9e194b0908bf05e0889b4c6e2a2304adcc980a5c3199a246d8c not found: ID does not exist" Nov 23 07:41:14 crc kubenswrapper[4559]: I1123 07:41:14.741710 4559 scope.go:117] "RemoveContainer" containerID="e5b2c31e701ac4238b9694f1e091432b2530370472ef8bce022425afd72c4acd" Nov 23 07:41:14 crc kubenswrapper[4559]: E1123 07:41:14.741928 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5b2c31e701ac4238b9694f1e091432b2530370472ef8bce022425afd72c4acd\": container with ID starting with e5b2c31e701ac4238b9694f1e091432b2530370472ef8bce022425afd72c4acd not found: ID does not exist" containerID="e5b2c31e701ac4238b9694f1e091432b2530370472ef8bce022425afd72c4acd" Nov 23 07:41:14 crc kubenswrapper[4559]: I1123 07:41:14.741956 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5b2c31e701ac4238b9694f1e091432b2530370472ef8bce022425afd72c4acd"} err="failed to get container status \"e5b2c31e701ac4238b9694f1e091432b2530370472ef8bce022425afd72c4acd\": rpc error: code = NotFound desc = could not find container \"e5b2c31e701ac4238b9694f1e091432b2530370472ef8bce022425afd72c4acd\": container with ID starting with e5b2c31e701ac4238b9694f1e091432b2530370472ef8bce022425afd72c4acd not found: ID does not exist" Nov 23 07:41:14 crc kubenswrapper[4559]: I1123 07:41:14.741973 4559 scope.go:117] "RemoveContainer" containerID="e922e7c9ad2fdbe32e0ad79f1babd55970e7cc403abbd4632fad7c43796f0474" Nov 23 07:41:14 crc kubenswrapper[4559]: E1123 07:41:14.742188 4559 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e922e7c9ad2fdbe32e0ad79f1babd55970e7cc403abbd4632fad7c43796f0474\": container with ID starting with e922e7c9ad2fdbe32e0ad79f1babd55970e7cc403abbd4632fad7c43796f0474 not found: ID does not exist" containerID="e922e7c9ad2fdbe32e0ad79f1babd55970e7cc403abbd4632fad7c43796f0474" Nov 23 07:41:14 crc kubenswrapper[4559]: I1123 07:41:14.742215 4559 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e922e7c9ad2fdbe32e0ad79f1babd55970e7cc403abbd4632fad7c43796f0474"} err="failed to get container status \"e922e7c9ad2fdbe32e0ad79f1babd55970e7cc403abbd4632fad7c43796f0474\": rpc error: code = NotFound desc = could not find container \"e922e7c9ad2fdbe32e0ad79f1babd55970e7cc403abbd4632fad7c43796f0474\": container with ID starting with e922e7c9ad2fdbe32e0ad79f1babd55970e7cc403abbd4632fad7c43796f0474 not found: ID does not exist" Nov 23 07:41:16 crc kubenswrapper[4559]: I1123 07:41:16.281032 4559 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8db86fb4-90eb-472b-8540-fa7d61dc3677" path="/var/lib/kubelet/pods/8db86fb4-90eb-472b-8540-fa7d61dc3677/volumes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515110535257024451 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015110535257017366 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015110526336016507 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015110526336015457 5ustar corecore